首页 > 最新文献

2021 RIVF International Conference on Computing and Communication Technologies (RIVF)最新文献

英文 中文
Multiple Imputation by Generative Adversarial Networks for Classification with Incomplete Data 生成对抗网络在不完全数据分类中的多重输入
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642138
Bao Ngoc Vi, Dinh Tan Nguyen, Cao Truong Tran, Huu Phuc Ngo, Chi Cong Nguyen, Hai-Hong Phan
Missing values present as the most common problem in real-world data science. Inadequate treatment of missing values could often result in mass errors. Hence missing values should be managed conscientiously for classification. Generative Adversarial Networks (GANs) have been applied for imputing missing values in most recent years. This paper proposes a multiple imputation method to estimate missing values for classification through the integration of GAN and ensemble learning. Our propose method MIGAN utilises GAN to generate different training observations which are then used to conduct ensemble classifiers for classification with missing data. We conducted our experiments examine MIGAN on various data sets as well as comparing MIGAN with the state-of-the-art imputation methods. The experimental results show significant results, which highlights the accuracy of MIGAN in classifying the missing data.
缺失值是现实世界数据科学中最常见的问题。对缺失值处理不当往往会导致大量误差。因此,缺失值应认真管理分类。近年来,生成对抗网络(GANs)被广泛应用于缺失值的估算。本文提出了一种将GAN和集成学习相结合的多重输入方法来估计分类中的缺失值。我们提出的方法MIGAN利用GAN生成不同的训练观测值,然后使用这些观测值进行集成分类器对缺失数据进行分类。我们在各种数据集上进行了实验,并将MIGAN与最先进的估算方法进行了比较。实验结果显示了显著的结果,表明了MIGAN对缺失数据分类的准确性。
{"title":"Multiple Imputation by Generative Adversarial Networks for Classification with Incomplete Data","authors":"Bao Ngoc Vi, Dinh Tan Nguyen, Cao Truong Tran, Huu Phuc Ngo, Chi Cong Nguyen, Hai-Hong Phan","doi":"10.1109/RIVF51545.2021.9642138","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642138","url":null,"abstract":"Missing values present as the most common problem in real-world data science. Inadequate treatment of missing values could often result in mass errors. Hence missing values should be managed conscientiously for classification. Generative Adversarial Networks (GANs) have been applied for imputing missing values in most recent years. This paper proposes a multiple imputation method to estimate missing values for classification through the integration of GAN and ensemble learning. Our propose method MIGAN utilises GAN to generate different training observations which are then used to conduct ensemble classifiers for classification with missing data. We conducted our experiments examine MIGAN on various data sets as well as comparing MIGAN with the state-of-the-art imputation methods. The experimental results show significant results, which highlights the accuracy of MIGAN in classifying the missing data.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"15 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73705059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
MC-OCR Challenge 2021: Simple approach for receipt information extraction and quality evaluation MC-OCR挑战2021:收据信息提取和质量评估的简单方法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642150
C. M. Nguyen, Vi Van Ngo, Dang Duy Nguyen
This challenge organized at the RIVF conference 2021 [12], with two tasks including (1) image quality assessment (IQA) of the captured receipt, and (2) key information extraction (KIE) of required fields, our team came up with a solution based on extracting image patches for task 1 and Yolov5 + VietOCR for task 2. Our solution achieved 0.149 of the RMSE score for task 1 (rank 7) and 0.219 of the CER score for task 2 (rank 1). Our code is available at https://github.com/cuongngm/RIVF2021.
这个挑战是在RIVF conference 2021上组织的[12],有两个任务,包括(1)捕获收据的图像质量评估(IQA)和(2)必要字段的关键信息提取(KIE),我们的团队提出了一个基于提取图像补丁的解决方案,用于任务1,Yolov5 + VietOCR用于任务2。我们的解决方案在任务1(排名7)的RMSE得分为0.149,在任务2(排名1)的CER得分为0.219。我们的代码可在https://github.com/cuongngm/RIVF2021上获得。
{"title":"MC-OCR Challenge 2021: Simple approach for receipt information extraction and quality evaluation","authors":"C. M. Nguyen, Vi Van Ngo, Dang Duy Nguyen","doi":"10.1109/RIVF51545.2021.9642150","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642150","url":null,"abstract":"This challenge organized at the RIVF conference 2021 [12], with two tasks including (1) image quality assessment (IQA) of the captured receipt, and (2) key information extraction (KIE) of required fields, our team came up with a solution based on extracting image patches for task 1 and Yolov5 + VietOCR for task 2. Our solution achieved 0.149 of the RMSE score for task 1 (rank 7) and 0.219 of the CER score for task 2 (rank 1). Our code is available at https://github.com/cuongngm/RIVF2021.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"86 1","pages":"1-4"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74199475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Real-time Multispectral Algorithm for Robust Pedestrian Detection 一种实时多光谱鲁棒行人检测算法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642066
Vu Hiep Dao, Hieu Mac, Duc Tran
Low light conditions are known to create a notable challenge to the applicability of deep learning in a wide variety of computer vision applications. In this paper, we develop a detection method for real-time multispectral pedestrians that fuses color image (i.e., red-green-blue or RBG) with thermal image to provide a reliable object vision. Such combination is achieved using the confidence scores that are computed based on the illumination measure of a given input image. We evaluate the proposed algorithm on KAIST dataset. Such method is observed to give a 34.11% Log Average Miss Rate, operate in real-time, and thus, being ready to deploy in practice.
众所周知,低光条件对深度学习在各种计算机视觉应用中的适用性构成了显著的挑战。在本文中,我们开发了一种实时多光谱行人检测方法,该方法将彩色图像(即红绿蓝或RBG)与热图像融合,以提供可靠的目标视觉。使用基于给定输入图像的照明度量计算的置信度分数来实现这种组合。我们在KAIST数据集上对该算法进行了评估。据观察,该方法的测井平均漏失率为34.11%,可以实时操作,因此可以在实践中部署。
{"title":"A Real-time Multispectral Algorithm for Robust Pedestrian Detection","authors":"Vu Hiep Dao, Hieu Mac, Duc Tran","doi":"10.1109/RIVF51545.2021.9642066","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642066","url":null,"abstract":"Low light conditions are known to create a notable challenge to the applicability of deep learning in a wide variety of computer vision applications. In this paper, we develop a detection method for real-time multispectral pedestrians that fuses color image (i.e., red-green-blue or RBG) with thermal image to provide a reliable object vision. Such combination is achieved using the confidence scores that are computed based on the illumination measure of a given input image. We evaluate the proposed algorithm on KAIST dataset. Such method is observed to give a 34.11% Log Average Miss Rate, operate in real-time, and thus, being ready to deploy in practice.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"18 4 1","pages":"1-4"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83455735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A GAN-based approach for password guessing 一种基于gan的密码猜测方法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642098
Bao Ngoc Vi, Nguyen Ngoc Tran, Trung Giap Vu The
Password is the most widely used authenticate method. Individuals ordinarily have numerous passwords for their documents or devices, and, in some cases, they need to recover them with password guessing tools. Most popular guessing tools require a dictionary of common passwords to check with password hashes. Thus, generative adversarial networks (GANs) are suitable choices to automatically create a high-quality dictionary without any additional information from experts or password structures. One of the successful GAN-based models is the PassGAN. However, existing GAN-based models suffer from the discrete nature of passwords. Therefore, we proposed and evaluated two improvement of the PassGAN model to tackle this problem: the GS-PassGAN model using Gumbel-Softmax relaxation and the S-PassGAN using a smooth representation of a real password obtained by an additional Auto-Encoder. Experiment results on three different popular datasets show that the proposed method is better than the PassGAN both in the standalone and combining cases. Moreover, the matching rate of the proposed method can be increased by more than 5%.
密码是使用最广泛的认证方法。个人的文档或设备通常有很多密码,在某些情况下,他们需要使用密码猜测工具来恢复密码。大多数流行的猜测工具需要一个常用密码字典来检查密码哈希值。因此,生成对抗网络(GANs)是自动创建高质量字典的合适选择,无需任何来自专家或密码结构的额外信息。其中一个成功的基于gan的模型是PassGAN。然而,现有的基于gan的模型受到密码离散性的影响。因此,我们提出并评估了两种PassGAN模型的改进来解决这个问题:使用Gumbel-Softmax松弛的GS-PassGAN模型和使用由额外的Auto-Encoder获得的真实密码的平滑表示的S-PassGAN模型。在三种不同的流行数据集上的实验结果表明,该方法在单独和组合情况下都优于PassGAN。此外,该方法的匹配率可提高5%以上。
{"title":"A GAN-based approach for password guessing","authors":"Bao Ngoc Vi, Nguyen Ngoc Tran, Trung Giap Vu The","doi":"10.1109/RIVF51545.2021.9642098","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642098","url":null,"abstract":"Password is the most widely used authenticate method. Individuals ordinarily have numerous passwords for their documents or devices, and, in some cases, they need to recover them with password guessing tools. Most popular guessing tools require a dictionary of common passwords to check with password hashes. Thus, generative adversarial networks (GANs) are suitable choices to automatically create a high-quality dictionary without any additional information from experts or password structures. One of the successful GAN-based models is the PassGAN. However, existing GAN-based models suffer from the discrete nature of passwords. Therefore, we proposed and evaluated two improvement of the PassGAN model to tackle this problem: the GS-PassGAN model using Gumbel-Softmax relaxation and the S-PassGAN using a smooth representation of a real password obtained by an additional Auto-Encoder. Experiment results on three different popular datasets show that the proposed method is better than the PassGAN both in the standalone and combining cases. Moreover, the matching rate of the proposed method can be increased by more than 5%.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"16 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90542453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Autoencoder-based Method for Targeted Attack on Deep Neural Network Models 基于自编码器的深度神经网络模型目标攻击方法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642102
D. Nguyen, Do Minh Kha, Pham Thi To Nga, Pham Ngoc Hung
This paper presents an autoencoder-based method for a targeted attack on deep neural network models, named AE4DNN. The proposed method aims to improve the existing targeted attacks in terms of their generalization, transferability, and the trade-off between the quality of adversarial examples and the computational cost. The idea of AE4DNN is that an autoencoder model is trained from a balanced subset of the training set. The trained autoencoder model is then used to generate adversarial examples from the remaining subset of the training set, produce adversarial examples from new samples, and attack other DNN models. To demonstrate the effectiveness of AE4DNN, the compared methods are box-constrained L-BFGS, Carlini-Wagner ‖L‖2 attack, and AAE. The comprehensive experiment on MNIST has shown that AE4DNN can gain a better transferability, improve generalization, and generate high quality of adversarial examples while requiring a low cost of computation. This initial result demonstrates the potential ability of AE4DNN in practice, which would help to reduce the effort of testing deep neural network models.
本文提出了一种基于自编码器的深度神经网络模型定向攻击方法,命名为AE4DNN。提出的方法旨在从泛化、可转移性以及对抗性示例的质量和计算成本之间的权衡等方面改进现有的目标攻击。AE4DNN的思想是从训练集的平衡子集中训练自编码器模型。然后使用训练好的自编码器模型从训练集的剩余子集中生成对抗性示例,从新样本中生成对抗性示例,并攻击其他DNN模型。为了证明AE4DNN的有效性,比较的方法是盒约束的L- bfgs, Carlini-Wagner‖L‖2攻击和AAE。在MNIST上的综合实验表明,AE4DNN可以获得更好的可转移性,提高泛化能力,生成高质量的对抗样例,同时需要较低的计算成本。这一初步结果证明了AE4DNN在实践中的潜在能力,这将有助于减少测试深度神经网络模型的工作量。
{"title":"An Autoencoder-based Method for Targeted Attack on Deep Neural Network Models","authors":"D. Nguyen, Do Minh Kha, Pham Thi To Nga, Pham Ngoc Hung","doi":"10.1109/RIVF51545.2021.9642102","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642102","url":null,"abstract":"This paper presents an autoencoder-based method for a targeted attack on deep neural network models, named AE4DNN. The proposed method aims to improve the existing targeted attacks in terms of their generalization, transferability, and the trade-off between the quality of adversarial examples and the computational cost. The idea of AE4DNN is that an autoencoder model is trained from a balanced subset of the training set. The trained autoencoder model is then used to generate adversarial examples from the remaining subset of the training set, produce adversarial examples from new samples, and attack other DNN models. To demonstrate the effectiveness of AE4DNN, the compared methods are box-constrained L-BFGS, Carlini-Wagner ‖L‖2 attack, and AAE. The comprehensive experiment on MNIST has shown that AE4DNN can gain a better transferability, improve generalization, and generate high quality of adversarial examples while requiring a low cost of computation. This initial result demonstrates the potential ability of AE4DNN in practice, which would help to reduce the effort of testing deep neural network models.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"8 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89185825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fuzzy C-Medoids Clustering Based on Interval Type-2 Inituitionistic Fuzzy Sets 基于区间2型初始模糊集的模糊c -介质聚类
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642067
Nguyễn Anh Cường, D. Mai, Do Viet Duc, Trong Hop Dang, L. Ngo, L. T. Pham
For clustering problems, each data sample has the potential to belong to many different clusters depending on the similarity. However, besides the degree of similarity and non-similarity, there is a degree of hesitation in determining whether or not a data sample belongs to a defined cluster. Besides the fuzzy c-means algorithm (FCM), another popular algorithm is fuzzy C-medoids clustering (FCMdd). FCMdd chooses several existing objects as the cluster centroids, while FCM considers the samples’ weighted average to be the cluster centroid. This subtle difference causes the FCMdd is more resistant to interference than FCM. Since noise samples will more easily affect the center of centroids of the FCM, it is easier to create clustering results with great accuracy. In this study, we proposed a method for extending the fuzzy c-medoids clustering based on interval type-2 intuitionistic fuzzy sets, named the interval type-2 intuitionistic fuzzy c-medoids clustering algorithm (IT2IFCMdd). With this combination, the proposed algorithm can take advantage of both the fuzzy c-medoids clustering (FCMdd) method and the interval type-2 intuitionistic fuzzy sets applied to the clustering problem. Experiments performed on data sets commonly used in machine learning show that the proposed method gives better clustering results in most experimental cases.
对于聚类问题,每个数据样本都有可能根据相似度属于许多不同的聚类。然而,除了相似度和非相似度之外,在确定数据样本是否属于已定义的聚类时还存在一定程度的犹豫。除了模糊c-均值算法(FCM)外,另一种流行的算法是模糊c-媒质聚类(FCMdd)。FCMdd选择几个现有的目标作为聚类质心,而FCM则将样本的加权平均值作为聚类质心。这种细微的差别使得FCMdd比FCM更能抵抗干扰。由于噪声样本更容易影响FCM的质心中心,因此更容易产生精度高的聚类结果。本文提出了一种基于区间2型直觉模糊集的模糊c-媒质聚类扩展方法,命名为区间2型直觉模糊c-媒质聚类算法(IT2IFCMdd)。该算法结合了模糊c-介质聚类(FCMdd)方法和区间2型直觉模糊集的优点,有效地解决了聚类问题。在机器学习中常用的数据集上进行的实验表明,在大多数实验情况下,所提出的方法具有更好的聚类结果。
{"title":"Fuzzy C-Medoids Clustering Based on Interval Type-2 Inituitionistic Fuzzy Sets","authors":"Nguyễn Anh Cường, D. Mai, Do Viet Duc, Trong Hop Dang, L. Ngo, L. T. Pham","doi":"10.1109/RIVF51545.2021.9642067","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642067","url":null,"abstract":"For clustering problems, each data sample has the potential to belong to many different clusters depending on the similarity. However, besides the degree of similarity and non-similarity, there is a degree of hesitation in determining whether or not a data sample belongs to a defined cluster. Besides the fuzzy c-means algorithm (FCM), another popular algorithm is fuzzy C-medoids clustering (FCMdd). FCMdd chooses several existing objects as the cluster centroids, while FCM considers the samples’ weighted average to be the cluster centroid. This subtle difference causes the FCMdd is more resistant to interference than FCM. Since noise samples will more easily affect the center of centroids of the FCM, it is easier to create clustering results with great accuracy. In this study, we proposed a method for extending the fuzzy c-medoids clustering based on interval type-2 intuitionistic fuzzy sets, named the interval type-2 intuitionistic fuzzy c-medoids clustering algorithm (IT2IFCMdd). With this combination, the proposed algorithm can take advantage of both the fuzzy c-medoids clustering (FCMdd) method and the interval type-2 intuitionistic fuzzy sets applied to the clustering problem. Experiments performed on data sets commonly used in machine learning show that the proposed method gives better clustering results in most experimental cases.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"11 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89561300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AG-ResUNet++: An Improved Encoder-Decoder Based Method for Polyp Segmentation in Colonoscopy Images ag - resunet++:一种改进的基于编码器-解码器的结肠镜图像息肉分割方法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642070
Nguyen Ba Hung, Thanh Duc Nguyen, Thai Van Chien, D. V. Sang
Colorectal cancer is one of the most prevalent causes of cancer-related death. Early polyp segmentation in colonoscopy is helpful in diagnosing and preventing colorectal cancer. However, this task a challenging due to variations in the appearance of polyps. This paper proposes a new encoder-decoder-based method called AG-ResUNet++ that leverages attention gate mechanism and residual connections to enhance the performance of the existing UNet++ model in the polyp segmentation task. Our method considerably outperforms other state-of-the-art methods on the popular polyp segmentation datasets, including KvasirSEG and CVC-612.
结直肠癌是癌症相关死亡的最普遍原因之一。结肠镜下早期息肉分割有助于结直肠癌的诊断和预防。然而,由于息肉外观的变化,这项任务具有挑战性。本文提出了一种新的基于编码器-解码器的方法ag - reun++,该方法利用注意门机制和剩余连接来提高现有un++模型在息肉分割任务中的性能。我们的方法在流行的息肉分割数据集上显著优于其他最先进的方法,包括KvasirSEG和CVC-612。
{"title":"AG-ResUNet++: An Improved Encoder-Decoder Based Method for Polyp Segmentation in Colonoscopy Images","authors":"Nguyen Ba Hung, Thanh Duc Nguyen, Thai Van Chien, D. V. Sang","doi":"10.1109/RIVF51545.2021.9642070","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642070","url":null,"abstract":"Colorectal cancer is one of the most prevalent causes of cancer-related death. Early polyp segmentation in colonoscopy is helpful in diagnosing and preventing colorectal cancer. However, this task a challenging due to variations in the appearance of polyps. This paper proposes a new encoder-decoder-based method called AG-ResUNet++ that leverages attention gate mechanism and residual connections to enhance the performance of the existing UNet++ model in the polyp segmentation task. Our method considerably outperforms other state-of-the-art methods on the popular polyp segmentation datasets, including KvasirSEG and CVC-612.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"14 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82405972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
MC-OCR Challenge: Mobile-Captured Image Document Recognition for Vietnamese Receipts MC-OCR挑战:越南收据的移动捕获图像文档识别
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642077
Xuan-Son Vu, Quang-Anh Bui, Nhu-Van Nguyen, Thi-Tuyet-Hai Nguyen, Thanh Vu
The paper describes the organisation of the "Mobile Captured Receipt Recognition Challenge" (MC-OCR) task at the RIVF conference 2021 1 on recognizing the fine-grained information in Vietnamese receipts captured using mobile devices. The task is organized as a multi-tasking model on a dataset containing 2,436 Vietnamese receipts. The participants were challenged to build a model that is capable of (1) predicting receipt’s quality based on readable information, and (2) recognizing textual information of four required information (i.e., "SELLER", "SELLER ADDRESS", "TIMESTAMP", and "TOTAL COST") in the receipts. MC-OCR challenge happened in one month and top winners of each task will present their solutions at RIVF 2021. Participants were competing on CodaLab.Org from 05th December 2020 to 23rd January 2021. All participants with valid submitted results were encouraged to submit their papers. Within one month, the challenge has attracted 105 participants and recorded about 1,285 submission entries.
该论文描述了在2021年RIVF会议1上组织的“移动捕获收据识别挑战”(MC-OCR)任务,该任务旨在识别使用移动设备捕获的越南收据中的细粒度信息。该任务在包含2,436个越南收据的数据集上组织为多任务模型。参与者被要求建立一个模型,该模型能够(1)基于可读信息预测收据的质量,(2)识别收据中四个必需信息(即“卖方”、“卖方地址”、“时间戳”和“总成本”)的文本信息。MC-OCR挑战在一个月内进行,每个任务的优胜者将在RIVF 2021上展示他们的解决方案。参与者在CodaLab上进行竞争。2020年12月5日至2021年1月23日。我们鼓励所有提交了有效结果的参与者提交论文。在一个月内,这项挑战吸引了105名参与者,并记录了约1,285份参赛作品。
{"title":"MC-OCR Challenge: Mobile-Captured Image Document Recognition for Vietnamese Receipts","authors":"Xuan-Son Vu, Quang-Anh Bui, Nhu-Van Nguyen, Thi-Tuyet-Hai Nguyen, Thanh Vu","doi":"10.1109/RIVF51545.2021.9642077","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642077","url":null,"abstract":"The paper describes the organisation of the \"Mobile Captured Receipt Recognition Challenge\" (MC-OCR) task at the RIVF conference 2021 1 on recognizing the fine-grained information in Vietnamese receipts captured using mobile devices. The task is organized as a multi-tasking model on a dataset containing 2,436 Vietnamese receipts. The participants were challenged to build a model that is capable of (1) predicting receipt’s quality based on readable information, and (2) recognizing textual information of four required information (i.e., \"SELLER\", \"SELLER ADDRESS\", \"TIMESTAMP\", and \"TOTAL COST\") in the receipts. MC-OCR challenge happened in one month and top winners of each task will present their solutions at RIVF 2021. Participants were competing on CodaLab.Org from 05th December 2020 to 23rd January 2021. All participants with valid submitted results were encouraged to submit their papers. Within one month, the challenge has attracted 105 participants and recorded about 1,285 submission entries.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"1 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79144459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Deep neural network based learning to rank for address standardization 基于深度神经网络学习的地址排序标准化
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642079
Hai Cao, Viet-Trung Tran
Address standardization is the process of converting and mapping free-form addresses into a standard structured format. For many business cases, the addresses are entered into the information systems by end-users. They are often noisy, uncompleted, and in different formatted styles. In this paper, we propose a deep learning-based approach to the address standardization challenge. Our key idea is to leverage a Siamese neural network model to embed raw inputs and standardized addresses into a single latent multi-dimensional space. Thus, the corresponding of the raw input address is the one with the highest-ranking score. Our experiments demonstrate that our best model achieved 95.41% accuracy, which is 6.6% improvement from the current state of the art.
地址标准化是将自由格式地址转换和映射为标准结构化格式的过程。对于许多业务案例,地址由最终用户输入到信息系统中。它们通常是嘈杂的、未完成的,并且格式风格不同。在本文中,我们提出了一种基于深度学习的方法来解决标准化挑战。我们的关键思想是利用暹罗神经网络模型将原始输入和标准化地址嵌入到单个潜在的多维空间中。因此,原始输入地址对应的是排名得分最高的地址。我们的实验表明,我们的最佳模型达到了95.41%的准确率,比目前的技术水平提高了6.6%。
{"title":"Deep neural network based learning to rank for address standardization","authors":"Hai Cao, Viet-Trung Tran","doi":"10.1109/RIVF51545.2021.9642079","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642079","url":null,"abstract":"Address standardization is the process of converting and mapping free-form addresses into a standard structured format. For many business cases, the addresses are entered into the information systems by end-users. They are often noisy, uncompleted, and in different formatted styles. In this paper, we propose a deep learning-based approach to the address standardization challenge. Our key idea is to leverage a Siamese neural network model to embed raw inputs and standardized addresses into a single latent multi-dimensional space. Thus, the corresponding of the raw input address is the one with the highest-ranking score. Our experiments demonstrate that our best model achieved 95.41% accuracy, which is 6.6% improvement from the current state of the art.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"11 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84177520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Hyperspectral Image Denoising Approach via Low-Rank Matrix Recovery and Greedy Bilateral 基于低秩矩阵恢复和贪心双边的高光谱图像去噪方法
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642145
Anh Tuan Vuong, Van Ha Tang, L. Ngo
The hyperspectral image (HSI) can provide useful information about the desired objects using spectral, spatial, and band channels. However, the image quality is typically distorted due to the limitations of sensing conditions and hardware operations. Consequently, the HSI is typically contaminated by a mixture noise during the acquisition process, including dead lines, stripes, Gaussian noise and impulse noise. In this paper, we introduce a new denoising model based on low-rank matrix recovery (LRMR), which can effectively remove various kinds of noise in HSI data. The low-rank property of the hyperspectral imagery is exploited by converting a patch of the HSI data from 3-D matrix into a 2-D matrix. The dead lines, stripes, and impulse noise are all modelled as sparse noise. To efficiently remove mixed noise and enhance performance, we develop an iterative algorithm using greedy bilateral technique to solve the optimization problem. To illustrate the proposed method’s efficacy in restoring HSI, both simulated and real-world HSI experiments are conducted.
高光谱图像(HSI)可以通过光谱、空间和波段通道提供有关目标的有用信息。然而,由于传感条件和硬件操作的限制,图像质量通常会失真。因此,在采集过程中,恒生指数通常受到混合噪声的污染,包括死线、条纹、高斯噪声和脉冲噪声。本文提出了一种新的基于低秩矩阵恢复(LRMR)的去噪模型,该模型能够有效地去除恒指数据中的各种噪声。利用高光谱图像的低秩特性,将一块HSI数据从3-D矩阵转换为2-D矩阵。死线、条纹和脉冲噪声都被建模为稀疏噪声。为了有效地去除混合噪声并提高性能,我们开发了一种使用贪心双边技术的迭代算法来解决优化问题。为了说明该方法在恢复HSI方面的有效性,进行了模拟和真实的HSI实验。
{"title":"A Hyperspectral Image Denoising Approach via Low-Rank Matrix Recovery and Greedy Bilateral","authors":"Anh Tuan Vuong, Van Ha Tang, L. Ngo","doi":"10.1109/RIVF51545.2021.9642145","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642145","url":null,"abstract":"The hyperspectral image (HSI) can provide useful information about the desired objects using spectral, spatial, and band channels. However, the image quality is typically distorted due to the limitations of sensing conditions and hardware operations. Consequently, the HSI is typically contaminated by a mixture noise during the acquisition process, including dead lines, stripes, Gaussian noise and impulse noise. In this paper, we introduce a new denoising model based on low-rank matrix recovery (LRMR), which can effectively remove various kinds of noise in HSI data. The low-rank property of the hyperspectral imagery is exploited by converting a patch of the HSI data from 3-D matrix into a 2-D matrix. The dead lines, stripes, and impulse noise are all modelled as sparse noise. To efficiently remove mixed noise and enhance performance, we develop an iterative algorithm using greedy bilateral technique to solve the optimization problem. To illustrate the proposed method’s efficacy in restoring HSI, both simulated and real-world HSI experiments are conducted.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"31 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88457171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 RIVF International Conference on Computing and Communication Technologies (RIVF)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1