Comparing the Robustness of Classical and Deep Learning Techniques for Text Classification

Quynh Tran, Krystsina Shpileuskaya, Elaine Zaunseder, Larissa Putzar, S. Blankenburg
{"title":"Comparing the Robustness of Classical and Deep Learning Techniques for Text Classification","authors":"Quynh Tran, Krystsina Shpileuskaya, Elaine Zaunseder, Larissa Putzar, S. Blankenburg","doi":"10.1109/IJCNN55064.2022.9892242","DOIUrl":null,"url":null,"abstract":"Deep learning algorithms achieve exceptional accuracies in various tasks. Despite this success, those models are known to be prone to errors, i.e. low in robustness, due to differences between training and production environment. One might assume that more model complexity translates directly to more robustness. Therefore, we compare simple, classical models (logistic regression, support vector machine) with complex deep learning techniques (convolutional neural networks, transformers) to provide novel insights into the robustness of machine learning systems. In our approach, we assess the robustness by developing and applying three realistic perturbations, mimicking scanning, typing, and speech recognition errors occurring in inputs for text classification tasks. Hence, we performed a thorough study analyzing the impact of different perturbations with variable strengths on character and word level. A noteworthy finding is that algorithms with low complexity can achieve high robustness. Additionally, we demonstrate that augmented training regarding a specific perturbation can strengthen the chosen models' robustness against other perturbations without reducing their accuracy. Our results can impact the selection of machine learning models and provide a guideline on how to examine the robustness of text classification methods for real-world applications. Moreover, our implementation is publicly available, which contributes to the development of more robust machine learning systems.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN55064.2022.9892242","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Deep learning algorithms achieve exceptional accuracies in various tasks. Despite this success, those models are known to be prone to errors, i.e. low in robustness, due to differences between training and production environment. One might assume that more model complexity translates directly to more robustness. Therefore, we compare simple, classical models (logistic regression, support vector machine) with complex deep learning techniques (convolutional neural networks, transformers) to provide novel insights into the robustness of machine learning systems. In our approach, we assess the robustness by developing and applying three realistic perturbations, mimicking scanning, typing, and speech recognition errors occurring in inputs for text classification tasks. Hence, we performed a thorough study analyzing the impact of different perturbations with variable strengths on character and word level. A noteworthy finding is that algorithms with low complexity can achieve high robustness. Additionally, we demonstrate that augmented training regarding a specific perturbation can strengthen the chosen models' robustness against other perturbations without reducing their accuracy. Our results can impact the selection of machine learning models and provide a guideline on how to examine the robustness of text classification methods for real-world applications. Moreover, our implementation is publicly available, which contributes to the development of more robust machine learning systems.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
比较经典和深度学习技术在文本分类中的鲁棒性
深度学习算法在各种任务中实现了卓越的准确性。尽管取得了成功,但由于训练和生产环境之间的差异,这些模型容易出错,即鲁棒性较低。有人可能会认为,更多的模型复杂性直接转化为更强的鲁棒性。因此,我们将简单的经典模型(逻辑回归、支持向量机)与复杂的深度学习技术(卷积神经网络、变压器)进行比较,以提供对机器学习系统鲁棒性的新见解。在我们的方法中,我们通过开发和应用三种现实的扰动来评估鲁棒性,模拟文本分类任务输入中出现的扫描、打字和语音识别错误。因此,我们进行了深入的研究,分析了不同强度的扰动对字符和单词水平的影响。一个值得注意的发现是,低复杂度的算法可以获得高鲁棒性。此外,我们证明了关于特定扰动的增强训练可以增强所选模型对其他扰动的鲁棒性,而不会降低其准确性。我们的研究结果可以影响机器学习模型的选择,并为如何检查现实世界应用中文本分类方法的鲁棒性提供指导。此外,我们的实现是公开的,这有助于开发更强大的机器学习系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Parameterization of Vector Symbolic Approach for Sequence Encoding Based Visual Place Recognition Nested compression of convolutional neural networks with Tucker-2 decomposition SQL-Rank++: A Novel Listwise Approach for Collaborative Ranking with Implicit Feedback ACTSS: Input Detection Defense against Backdoor Attacks via Activation Subset Scanning ADV-ResNet: Residual Network with Controlled Adversarial Regularization for Effective Classification of Practical Time Series Under Training Data Scarcity Problem
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1