Evaluation of Deep Learning Context-Sensitive Visualization Models

A. Dunn, D. Inkpen, Razvan Andonie
{"title":"Evaluation of Deep Learning Context-Sensitive Visualization Models","authors":"A. Dunn, D. Inkpen, Razvan Andonie","doi":"10.1109/IV56949.2022.00066","DOIUrl":null,"url":null,"abstract":"The introduction of Transformer neural networks has changed the landscape of Natural Language Processing (NLP) during the recent years. These models are very complex, and therefore hard to debug and explain. In this context, visual explanation became an attractive approach. The visualization of the path that leads to certain outputs of a model is at the core of visual explanation, as this illuminates the features or parts of the model that may need to be changed to achieve the desired results. In particular, one goal of a NLP visual explanation is to highlight the most significant parts of the text that have the greatest impact on the model output. Several visual explanation methods for NLP models were recently proposed. A major challenge is how to compare the performances of such methods since we cannot simply use the usual classification accuracy measures to evaluate the quality of visualizations. We need good metrics and rigorous criteria to measure how useful the extracted knowledge is for explaining the models. In addition, we want to visualize the differences between the knowledge extracted by different models, in order to be able to rank them. In this paper, we investigate how to evaluate explanations/visualizations resulted from machine learning models for text classification. The goal is not to improve the accuracy of a particular NLP classifier, but to assess the quality of the visualizations that explain its decisions. We describe several methods for evaluating the quality of NLP visualizations, including both automated techniques based on quantifiable measures and subjective techniques based on human judgements.","PeriodicalId":153161,"journal":{"name":"2022 26th International Conference Information Visualisation (IV)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 26th International Conference Information Visualisation (IV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IV56949.2022.00066","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The introduction of Transformer neural networks has changed the landscape of Natural Language Processing (NLP) during the recent years. These models are very complex, and therefore hard to debug and explain. In this context, visual explanation became an attractive approach. The visualization of the path that leads to certain outputs of a model is at the core of visual explanation, as this illuminates the features or parts of the model that may need to be changed to achieve the desired results. In particular, one goal of a NLP visual explanation is to highlight the most significant parts of the text that have the greatest impact on the model output. Several visual explanation methods for NLP models were recently proposed. A major challenge is how to compare the performances of such methods since we cannot simply use the usual classification accuracy measures to evaluate the quality of visualizations. We need good metrics and rigorous criteria to measure how useful the extracted knowledge is for explaining the models. In addition, we want to visualize the differences between the knowledge extracted by different models, in order to be able to rank them. In this paper, we investigate how to evaluate explanations/visualizations resulted from machine learning models for text classification. The goal is not to improve the accuracy of a particular NLP classifier, but to assess the quality of the visualizations that explain its decisions. We describe several methods for evaluating the quality of NLP visualizations, including both automated techniques based on quantifiable measures and subjective techniques based on human judgements.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
深度学习上下文敏感可视化模型的评价
近年来,变压器神经网络的引入改变了自然语言处理(NLP)的格局。这些模型非常复杂,因此很难调试和解释。在这种情况下,视觉解释成为一种有吸引力的方法。导致模型的某些输出的路径的可视化是可视化解释的核心,因为它阐明了可能需要更改以实现预期结果的模型的特征或部分。特别是,NLP可视化解释的一个目标是突出文本中对模型输出影响最大的最重要部分。最近提出了几种NLP模型的可视化解释方法。一个主要的挑战是如何比较这些方法的性能,因为我们不能简单地使用通常的分类精度度量来评估可视化的质量。我们需要良好的度量和严格的标准来衡量提取的知识对解释模型的有用程度。此外,我们希望可视化不同模型提取的知识之间的差异,以便能够对它们进行排序。在本文中,我们研究了如何评估机器学习模型对文本分类的解释/可视化结果。目标不是提高特定NLP分类器的准确性,而是评估解释其决策的可视化的质量。我们描述了几种评估NLP可视化质量的方法,包括基于可量化测量的自动化技术和基于人类判断的主观技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Phrase Features in Essay Report Sentences for Developing Critical Thinking Ability in a Fully Online Course Preoperative Image Segmentation for Organ Visualization Using Augmented Reality Technology During Open Liver Surgery Data. Information and Knowledge Visualization for Frequent Patterns VRGrid: Efficient Transformation of 2D Data into Pixel Grid Layout Augmenting the Reality of Situated Visualization
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1