Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform

Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto
{"title":"Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform","authors":"Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto","doi":"10.1109/ICITEE56407.2022.9954068","DOIUrl":null,"url":null,"abstract":"Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.","PeriodicalId":246279,"journal":{"name":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITEE56407.2022.9954068","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
雅加达超级应用平台自动收据识别的知识蒸馏
计算机视觉研究已经应用于日常应用,如艺术、社交媒体应用程序过滤器和人脸识别。这种出现是因为深度学习方法在计算机视觉领域的应用。深度学习研究提高了各种应用的服务质量。从推荐到检测系统现在依赖于深度学习模型。然而,目前许多模型对计算处理和存储空间的要求很高。在嵌入式设备或智能手机上使用有限的资源实现如此广泛的网络变得更具挑战性。在本研究中,我们着重于利用知识蒸馏方法开发一个计算资源少、精度高的模型。我们在从Badan Pendapatan Daerah、CORD和Kaggle数据集收集的收据和非收据图像的公共和私人数据集上评估我们的模型。之后,我们将其与常规卷积神经网络(CNN)和预训练模型进行比较。我们发现知识蒸馏分别只使用了CNN和预训练模型总权重的12%和5%。因此,我们看到了一种可能性,即知识蒸馏说明了作为一种可以在雅加达超级应用程序中实现自动收据识别的方法的潜在结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Measuring Trust Perception during Information Evaluation Comparative Study of Multi-channel and Omni-channel on Supply Chain Management Classifying Stress Mental State by using Power Spectral Density of Electroencephalography (EEG) Development of CNN Pruning Method for Earthquake Signal Imagery Classification Object Selection Using LSTM Networks for Spontaneous Gaze-Based Interaction
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1