The Best Parameter Tuning on RNN Layers for Indonesian Text Classification

Awaliyatul Hikmah, Sumarni Adi, Mulia Sulistiyono
{"title":"The Best Parameter Tuning on RNN Layers for Indonesian Text Classification","authors":"Awaliyatul Hikmah, Sumarni Adi, Mulia Sulistiyono","doi":"10.1109/ISRITI51436.2020.9315425","DOIUrl":null,"url":null,"abstract":"Recurrent Neural Network (RNN) is a deep learning architecture commonly used to process time series and sequence data. Various architectures have been developed to improve the performance of the algorithm in terms of both accuracy and computation time. Besides, the use of appropriate parameter values when building a neural network model also plays an important role in the quality and the outcome of the learning model. In this study, the model trained using RNN-Vanilla, LSTM, and GRU each with 4 different combinations of parameter settings, namely bidirectional mode (True, False), the number of neuron units on each layer (64, 128, 256), the number of RNN layers on the neural network (1, 2, 3), and the batch size when training the model (32, 64, 128). By combining all the parameter values, 162 trials were carried out to perform the task of classifying Indonesian language customer support tickets with four category classes. This study gives the result that the same network architecture but with different parameter combinations results in significant differences in the level of accuracy. The lowest accuracy of all experiments was 32.874% and the highest accuracy resulted was 84.369%. Overall, by calculating the average accuracy of each parameter value, the results obtained are: GRU has the best performance, accuracy tends to increase by activating bidirectional mode, increasing the number of neuron units in the hidden layer, and reducing the batch size. Meanwhile, the addition of the number of RNN layers on the neural network has no impact on increasing the level of accuracy.","PeriodicalId":325920,"journal":{"name":"2020 3rd International Seminar on Research of Information Technology and Intelligent Systems (ISRITI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2020-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 3rd International Seminar on Research of Information Technology and Intelligent Systems (ISRITI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISRITI51436.2020.9315425","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Recurrent Neural Network (RNN) is a deep learning architecture commonly used to process time series and sequence data. Various architectures have been developed to improve the performance of the algorithm in terms of both accuracy and computation time. Besides, the use of appropriate parameter values when building a neural network model also plays an important role in the quality and the outcome of the learning model. In this study, the model trained using RNN-Vanilla, LSTM, and GRU each with 4 different combinations of parameter settings, namely bidirectional mode (True, False), the number of neuron units on each layer (64, 128, 256), the number of RNN layers on the neural network (1, 2, 3), and the batch size when training the model (32, 64, 128). By combining all the parameter values, 162 trials were carried out to perform the task of classifying Indonesian language customer support tickets with four category classes. This study gives the result that the same network architecture but with different parameter combinations results in significant differences in the level of accuracy. The lowest accuracy of all experiments was 32.874% and the highest accuracy resulted was 84.369%. Overall, by calculating the average accuracy of each parameter value, the results obtained are: GRU has the best performance, accuracy tends to increase by activating bidirectional mode, increasing the number of neuron units in the hidden layer, and reducing the batch size. Meanwhile, the addition of the number of RNN layers on the neural network has no impact on increasing the level of accuracy.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
印尼语文本分类RNN层的最佳参数调优
递归神经网络(RNN)是一种深度学习架构,通常用于处理时间序列和序列数据。为了提高算法在精度和计算时间方面的性能,已经开发了各种体系结构。此外,在构建神经网络模型时,使用合适的参数值对学习模型的质量和结果也起着重要的作用。在本研究中,使用RNN- vanilla、LSTM和GRU训练的模型各有4种不同的参数设置组合,即双向模式(True, False)、每层神经元单元数(64、128、256)、神经网络上的RNN层数(1、2、3)和训练模型时的批大小(32、64、128)。通过组合所有参数值,进行了162次试验,以执行将印尼语客户支持票分为四类的任务。研究结果表明,在相同的网络结构下,不同的参数组合会导致准确率水平的显著差异。所有实验的最低准确率为32.874%,最高准确率为84.369%。总体而言,通过计算各参数值的平均准确率,得到的结果是:GRU的性能最好,激活双向模式、增加隐藏层神经元单元数、减小批处理大小,准确率有提高的趋势。同时,在神经网络上增加RNN层数对提高准确率水平没有影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Combined Firefly Algorithm-Random Forest to Classify Autistic Spectrum Disorders Analysis of Indonesia's Internet Topology Borders at the Autonomous System Level Influence Distribution Training Data on Performance Supervised Machine Learning Algorithms Design of Optimal Satellite Constellation for Indonesian Regional Navigation System based on GEO and GSO Satellites Real-time Testing on Improved Data Transmission Security in the Industrial Control System
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1