Pub Date : 2024-04-17DOI: 10.1177/14738716241238476
Merel de Leeuw den Bouter, Javier Lloret Pardo, Zeno Geradts, Marcel Worring
In high-stakes settings, Machine Learning models that can provide predictions that are interpretable for humans are crucial. This is even more true with the advent of complex deep learning based models with a huge number of tunable parameters. Recently, prototype-based methods have emerged as a promising approach to make deep learning interpretable. We particularly focus on the analysis of deepfake videos in a forensics context. Although prototype-based methods have been introduced for the detection of deepfake videos, their use in real-world scenarios still presents major challenges, in that prototypes tend to be overly similar and interpretability varies between prototypes. This paper proposes a Visual Analytics process model for prototype learning, and, based on this, presents ProtoExplorer, a Visual Analytics system for the exploration and refinement of prototype-based deepfake detection models. ProtoExplorer offers tools for visualizing and temporally filtering prototype-based predictions when working with video data. It disentangles the complexity of working with spatio-temporal prototypes, facilitating their visualization. It further enables the refinement of models by interactively deleting and replacing prototypes with the aim to achieve more interpretable and less biased predictions while preserving detection accuracy. The system was designed with forensic experts and evaluated in a number of rounds based on both open-ended think aloud evaluation and interviews. These sessions have confirmed the strength of our prototype-based exploration of deepfake videos while they provided the feedback needed to continuously improve the system.
{"title":"ProtoExplorer: Interpretable forensic analysis of deepfake videos using prototype exploration and refinement","authors":"Merel de Leeuw den Bouter, Javier Lloret Pardo, Zeno Geradts, Marcel Worring","doi":"10.1177/14738716241238476","DOIUrl":"https://doi.org/10.1177/14738716241238476","url":null,"abstract":"In high-stakes settings, Machine Learning models that can provide predictions that are interpretable for humans are crucial. This is even more true with the advent of complex deep learning based models with a huge number of tunable parameters. Recently, prototype-based methods have emerged as a promising approach to make deep learning interpretable. We particularly focus on the analysis of deepfake videos in a forensics context. Although prototype-based methods have been introduced for the detection of deepfake videos, their use in real-world scenarios still presents major challenges, in that prototypes tend to be overly similar and interpretability varies between prototypes. This paper proposes a Visual Analytics process model for prototype learning, and, based on this, presents ProtoExplorer, a Visual Analytics system for the exploration and refinement of prototype-based deepfake detection models. ProtoExplorer offers tools for visualizing and temporally filtering prototype-based predictions when working with video data. It disentangles the complexity of working with spatio-temporal prototypes, facilitating their visualization. It further enables the refinement of models by interactively deleting and replacing prototypes with the aim to achieve more interpretable and less biased predictions while preserving detection accuracy. The system was designed with forensic experts and evaluated in a number of rounds based on both open-ended think aloud evaluation and interviews. These sessions have confirmed the strength of our prototype-based exploration of deepfake videos while they provided the feedback needed to continuously improve the system.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"183 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140612431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-08DOI: 10.1177/14738716241239619
Raissa dos Santos Vieira, Hugo Alexandre Dantas do Nascimento, Joelma de Moura Ferreira, Les Foulds
Edge bundling is a technique used to improve the readability of large graph drawings by grouping edges to reduce visual complexity. This paper treats this task as a clustering problem, using compatibility metrics to evaluate solutions in an optimization pipeline combined with a clustering ensemble approach. The aim is to present the Clustering Ensemble-based Edge Bundling (CEBEB) method for solving the General-based Edge Bundling (GBEB) problem and report results for some given graphs. The CEBEB method proved very promising and generated better solutions than an existing evolutionary algorithm. Additionally, the paper introduces a new ensemble algorithm, specific for the GBEB, and reviews some previous results.
{"title":"Enhancing graph drawings through edge bundling using clustering ensembles","authors":"Raissa dos Santos Vieira, Hugo Alexandre Dantas do Nascimento, Joelma de Moura Ferreira, Les Foulds","doi":"10.1177/14738716241239619","DOIUrl":"https://doi.org/10.1177/14738716241239619","url":null,"abstract":"Edge bundling is a technique used to improve the readability of large graph drawings by grouping edges to reduce visual complexity. This paper treats this task as a clustering problem, using compatibility metrics to evaluate solutions in an optimization pipeline combined with a clustering ensemble approach. The aim is to present the Clustering Ensemble-based Edge Bundling (CEBEB) method for solving the General-based Edge Bundling (GBEB) problem and report results for some given graphs. The CEBEB method proved very promising and generated better solutions than an existing evolutionary algorithm. Additionally, the paper introduces a new ensemble algorithm, specific for the GBEB, and reviews some previous results.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"75 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140602315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-08DOI: 10.1177/14738716241240156
Luoxuan Weng, Shi Liu, Hang Zhu, Jiashun Sun, Wong Kam-Kwai, Dongming Han, Minfeng Zhu, Wei Chen
Large language models (LLMs) have gained popularity in various fields for their exceptional capability of generating human-like text. Their potential misuse has raised social concerns about plagiarism in academic contexts. However, effective artificial scientific text detection is a non-trivial task due to several challenges, including (1) the lack of a clear understanding of the differences between machine-generated and human-written scientific text, (2) the poor generalization performance of existing methods caused by out-of-distribution issues, and (3) the limited support for human-machine collaboration with sufficient interpretability during the detection process. In this paper, we first identify the critical distinctions between machine-generated and human-written scientific text through a quantitative experiment. Then, we propose a mixed-initiative workflow that combines human experts’ prior knowledge with machine intelligence, along with a visual analytics system to facilitate efficient and trustworthy scientific text detection. Finally, we demonstrate the effectiveness of our approach through two case studies and a controlled user study. We also provide design implications for interactive artificial text detection tools in high-stakes decision-making scenarios.
{"title":"Towards an understanding and explanation for mixed-initiative artificial scientific text detection","authors":"Luoxuan Weng, Shi Liu, Hang Zhu, Jiashun Sun, Wong Kam-Kwai, Dongming Han, Minfeng Zhu, Wei Chen","doi":"10.1177/14738716241240156","DOIUrl":"https://doi.org/10.1177/14738716241240156","url":null,"abstract":"Large language models (LLMs) have gained popularity in various fields for their exceptional capability of generating human-like text. Their potential misuse has raised social concerns about plagiarism in academic contexts. However, effective artificial scientific text detection is a non-trivial task due to several challenges, including (1) the lack of a clear understanding of the differences between machine-generated and human-written scientific text, (2) the poor generalization performance of existing methods caused by out-of-distribution issues, and (3) the limited support for human-machine collaboration with sufficient interpretability during the detection process. In this paper, we first identify the critical distinctions between machine-generated and human-written scientific text through a quantitative experiment. Then, we propose a mixed-initiative workflow that combines human experts’ prior knowledge with machine intelligence, along with a visual analytics system to facilitate efficient and trustworthy scientific text detection. Finally, we demonstrate the effectiveness of our approach through two case studies and a controlled user study. We also provide design implications for interactive artificial text detection tools in high-stakes decision-making scenarios.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"1 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140586471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-05DOI: 10.1177/14738716241239018
Alexander Strang, David Sewell, Alexander Kim, Kevin Alcedo, David Rosenbluth
How are the advantage relations between a set of agents playing a game organized and how do they reflect the structure of the game? In this paper, we illustrate ‘Principal Trade-off Analysis’ (PTA), a decomposition method that embeds games into a low-dimensional feature space. We argue that the embeddings are more revealing than previously demonstrated by developing an analogy to Principal Component Analysis (PCA). PTA represents an arbitrary two-player zero-sum game as linear combination of simple games via the projection of policy profiles into orthogonal 2D feature planes. We show that the feature planes represent unique strategic trade-offs and truncation of the sequence provides insightful model reduction and visualization. We demonstrate the validity of PTA on a quartet of games (Kuhn poker, RPS + 2, Blotto and Pokemon). In Kuhn poker, PTA clearly identifies the trade-off between bluffing and calling. In Blotto, PTA identifies game symmetries and specifies strategic trade-offs associated with distinct win conditions. These symmetries reveal limitations of PTA unaddressed in previous work. For Pokemon, PTA recovers clusters that naturally correspond to Pokemon types, correctly identifies the designed trade-off between those types, and discovers a rock-paper-scissor (RPS) cycle in the Pokemon generation type – all absent any specific information except game outcomes.
{"title":"Principal trade-off analysis","authors":"Alexander Strang, David Sewell, Alexander Kim, Kevin Alcedo, David Rosenbluth","doi":"10.1177/14738716241239018","DOIUrl":"https://doi.org/10.1177/14738716241239018","url":null,"abstract":"How are the advantage relations between a set of agents playing a game organized and how do they reflect the structure of the game? In this paper, we illustrate ‘Principal Trade-off Analysis’ (PTA), a decomposition method that embeds games into a low-dimensional feature space. We argue that the embeddings are more revealing than previously demonstrated by developing an analogy to Principal Component Analysis (PCA). PTA represents an arbitrary two-player zero-sum game as linear combination of simple games via the projection of policy profiles into orthogonal 2D feature planes. We show that the feature planes represent unique strategic trade-offs and truncation of the sequence provides insightful model reduction and visualization. We demonstrate the validity of PTA on a quartet of games (Kuhn poker, RPS + 2, Blotto and Pokemon). In Kuhn poker, PTA clearly identifies the trade-off between bluffing and calling. In Blotto, PTA identifies game symmetries and specifies strategic trade-offs associated with distinct win conditions. These symmetries reveal limitations of PTA unaddressed in previous work. For Pokemon, PTA recovers clusters that naturally correspond to Pokemon types, correctly identifies the designed trade-off between those types, and discovers a rock-paper-scissor (RPS) cycle in the Pokemon generation type – all absent any specific information except game outcomes.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"48 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140586479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-28DOI: 10.1177/14738716241236188
Maria Skeppstedt, Magnus Ahltorp, Kostiantyn Kucher, Matts Lindström
Word Rain is a development of the classic word cloud. It addresses some of the limitations of word clouds, in particular the lack of a semantically motivated positioning of the words, and the use of font size as a sole indicator of word prominence. Word Rain uses the semantic information encoded in a distributional semantics-based language model – reduced into one dimension – to position the words along the x-axis. Thereby, the horizontal positioning of the words reflects semantic similarity. Font size is still used to signal word prominence, but this signal is supplemented with a bar chart, as well as with the position of the words on the y-axis. We exemplify the use of Word Rain by three concrete visualization tasks, applied on different real-world texts and document collections on climate change. In these case studies, word2vec models, reduced to one dimension with t-SNE, are used to encode semantic similarity, and TF-IDF is used for measuring word prominence. We evaluate the technique further by carrying out domain expert reviews.
词雨是对经典词云的发展。它解决了词云的一些局限性问题,特别是缺乏以语义为基础的词定位,以及将字体大小作为衡量词突出度的唯一指标。字雨使用基于分布语义的语言模型中编码的语义信息(简化为一个维度)来沿 x 轴定位词语。因此,词语的水平定位反映了语义的相似性。字体大小仍用于表示单词的显著性,但这一信号通过条形图以及单词在 y 轴上的位置得到了补充。我们通过三个具体的可视化任务来示范 "词雨 "的使用,这些任务应用于不同的真实文本和有关气候变化的文档集。在这些案例研究中,用 t-SNE 将 word2vec 模型缩减到一个维度来编码语义相似性,用 TF-IDF 来测量词的显著性。我们通过进行领域专家评审来进一步评估该技术。
{"title":"From word clouds to Word Rain: Revisiting the classic word cloud to visualize climate change texts","authors":"Maria Skeppstedt, Magnus Ahltorp, Kostiantyn Kucher, Matts Lindström","doi":"10.1177/14738716241236188","DOIUrl":"https://doi.org/10.1177/14738716241236188","url":null,"abstract":"Word Rain is a development of the classic word cloud. It addresses some of the limitations of word clouds, in particular the lack of a semantically motivated positioning of the words, and the use of font size as a sole indicator of word prominence. Word Rain uses the semantic information encoded in a distributional semantics-based language model – reduced into one dimension – to position the words along the x-axis. Thereby, the horizontal positioning of the words reflects semantic similarity. Font size is still used to signal word prominence, but this signal is supplemented with a bar chart, as well as with the position of the words on the y-axis. We exemplify the use of Word Rain by three concrete visualization tasks, applied on different real-world texts and document collections on climate change. In these case studies, word2vec models, reduced to one dimension with t-SNE, are used to encode semantic similarity, and TF-IDF is used for measuring word prominence. We evaluate the technique further by carrying out domain expert reviews.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"9 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140322488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-07DOI: 10.1177/14738716241229437
Arran Zeyu Wang, David Borland, David Gotz
Counterfactuals – expressing what might have been true under different circumstances – have been widely applied in statistics and machine learning to help understand causal relationships. More recently, counterfactuals have begun to emerge as a technique being applied within visualization research. However, it remains unclear to what extent counterfactuals can aid with visual data communication. In this paper, we primarily focus on assessing the quality of users’ understanding of data when provided with counterfactual visualizations. We propose a preliminary model of causality comprehension by connecting theories from causal inference and visual data communication. Leveraging this model, we conducted an empirical study to explore how counterfactuals can improve users’ understanding of data in static visualizations. Our results indicate that visualizing counterfactuals had a positive impact on participants’ interpretations of causal relations within datasets. These results motivate a discussion of how to more effectively incorporate counterfactuals into data visualizations.
{"title":"An empirical study of counterfactual visualization to support visual causal inference","authors":"Arran Zeyu Wang, David Borland, David Gotz","doi":"10.1177/14738716241229437","DOIUrl":"https://doi.org/10.1177/14738716241229437","url":null,"abstract":"Counterfactuals – expressing what might have been true under different circumstances – have been widely applied in statistics and machine learning to help understand causal relationships. More recently, counterfactuals have begun to emerge as a technique being applied within visualization research. However, it remains unclear to what extent counterfactuals can aid with visual data communication. In this paper, we primarily focus on assessing the quality of users’ understanding of data when provided with counterfactual visualizations. We propose a preliminary model of causality comprehension by connecting theories from causal inference and visual data communication. Leveraging this model, we conducted an empirical study to explore how counterfactuals can improve users’ understanding of data in static visualizations. Our results indicate that visualizing counterfactuals had a positive impact on participants’ interpretations of causal relations within datasets. These results motivate a discussion of how to more effectively incorporate counterfactuals into data visualizations.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"95 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139956369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-30DOI: 10.1177/14738716231220536
Mostafa M Hamza, Ehsan Ullah, Abdelkader Baggag, Halima Bensmail, Michael Sedlmair, Michael Aupetit
Visual quality measures (VQMs) are designed to support analysts by automatically detecting and quantifying patterns in visualizations. We propose a new VQM for visual grouping patterns in scatterplots, called ClustML, which is trained on previously collected human subject judgments. Our model encodes scatterplots in the parametric space of a Gaussian Mixture Model and uses a classifier trained on human judgment data to estimate the perceptual complexity of grouping patterns. The numbers of initial mixture components and final combined groups quantify visual cluster patterns in scatterplots. It improves on existing VQMs, first, by better estimating human judgments on two-Gaussian cluster patterns and, second, by giving higher accuracy when ranking general cluster patterns in scatterplots. We use it to analyze kinship data for genome-wide association studies, in which experts rely on the visual analysis of large sets of scatterplots. We make the benchmark datasets and the new VQM available for practical use and further improvements.
{"title":"ClustML: A measure of cluster pattern complexity in scatterplots learnt from human-labeled groupings","authors":"Mostafa M Hamza, Ehsan Ullah, Abdelkader Baggag, Halima Bensmail, Michael Sedlmair, Michael Aupetit","doi":"10.1177/14738716231220536","DOIUrl":"https://doi.org/10.1177/14738716231220536","url":null,"abstract":"Visual quality measures (VQMs) are designed to support analysts by automatically detecting and quantifying patterns in visualizations. We propose a new VQM for visual grouping patterns in scatterplots, called ClustML, which is trained on previously collected human subject judgments. Our model encodes scatterplots in the parametric space of a Gaussian Mixture Model and uses a classifier trained on human judgment data to estimate the perceptual complexity of grouping patterns. The numbers of initial mixture components and final combined groups quantify visual cluster patterns in scatterplots. It improves on existing VQMs, first, by better estimating human judgments on two-Gaussian cluster patterns and, second, by giving higher accuracy when ranking general cluster patterns in scatterplots. We use it to analyze kinship data for genome-wide association studies, in which experts rely on the visual analysis of large sets of scatterplots. We make the benchmark datasets and the new VQM available for practical use and further improvements.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"163 1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2024-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139950084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-29DOI: 10.1177/14738716231216030
V. Ciorna, Guy Melançon, F. Petry, Mohammad Ghoniem
Virtual prototyping is increasingly used by businesses to streamline operations, cut costs, and enhance daily operations. This often includes a variety of modeling techniques among which, complex, black-box models. The path from model development to utilization in applied contexts is yet long. Domain experts need to be convinced of the validity of the models and to trust their predictions. To be used in the field, model capabilities need to be affordable, that is, allow rapid and interactive scenario building, even for non-experts. Complex relations governed by statistical interactions must be unveiled for users to understand unexpected predictions. We propose Interact, a model-agnostic, visual what-if tool for regression problems, supporting (1) the visualization of statistical interactions between features, (2) the creation of interactive what-if scenarios using predictive models, (3) the evaluation of model quality and building trust, and (4) the externalization of knowledge through model explainability. While the approach applies in various industrial contexts, we validate the application purpose and design with a detailed case study and a qualitative user study with engineers in the tire industry. By unraveling statistical interactions between features, the INTERACT tool proves to be useful to increase the transparency of black-box machine learning models. We also reflect on lessons learned concerning the development of visual what-if tools for virtual product development and beyond.
{"title":"Interact: A visual what-if analysis tool for virtual product design","authors":"V. Ciorna, Guy Melançon, F. Petry, Mohammad Ghoniem","doi":"10.1177/14738716231216030","DOIUrl":"https://doi.org/10.1177/14738716231216030","url":null,"abstract":"Virtual prototyping is increasingly used by businesses to streamline operations, cut costs, and enhance daily operations. This often includes a variety of modeling techniques among which, complex, black-box models. The path from model development to utilization in applied contexts is yet long. Domain experts need to be convinced of the validity of the models and to trust their predictions. To be used in the field, model capabilities need to be affordable, that is, allow rapid and interactive scenario building, even for non-experts. Complex relations governed by statistical interactions must be unveiled for users to understand unexpected predictions. We propose Interact, a model-agnostic, visual what-if tool for regression problems, supporting (1) the visualization of statistical interactions between features, (2) the creation of interactive what-if scenarios using predictive models, (3) the evaluation of model quality and building trust, and (4) the externalization of knowledge through model explainability. While the approach applies in various industrial contexts, we validate the application purpose and design with a detailed case study and a qualitative user study with engineers in the tire industry. By unraveling statistical interactions between features, the INTERACT tool proves to be useful to increase the transparency of black-box machine learning models. We also reflect on lessons learned concerning the development of visual what-if tools for virtual product development and beyond.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":" 31","pages":""},"PeriodicalIF":2.3,"publicationDate":"2023-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139143626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In recent times, visual analytics systems (VAS) have been used to solve various complex issues in diverse application domains. Nonetheless, an inherent drawback arises from the insufficient evaluation of VAS, resulting in occasional inaccuracies when it comes to analytical reasoning, information synthesis, and deriving insights from vast, ever-changing, ambiguous, and frequently contradictory data. Hence, the significance of implementing an appropriate evaluation methodology cannot be overstated, as it plays a pivotal role in enhancing the design and development of visualization systems. This paper assesses visualization systems by providing a systematic exploration of various evaluation strategies (ES). While several existing studies have examined some ES, the extent of comprehensive and systematic review for visualization research remains limited. In this work, we introduce seven state-of-the-art and widely recognized ES namely (1) dashboard comparison; (2) insight-based evaluation; (3) log data analysis; (4) Likert scales; (5) qualitative and quantitative analysis; (6) Nielsen’s heuristics; and (7) eye trackers. Moreover, it delves into their historical context and explores numerous applications where these ES have been employed, shedding light on the associated evaluation practices. Through our comprehensive review, we overview and analyze the predominant evaluation goals within the visualization community, elucidating their evolution and the inherent contrasts. Additionally, we identify the open challenges that arise with the emergence of new ES, while also highlighting the key themes gleaned from the existing literature that hold potential for further exploration in future studies.
近年来,可视分析系统(VAS)已被用于解决不同应用领域中的各种复杂问题。然而,可视分析系统固有的缺点是评估不足,导致在分析推理、信息合成以及从大量、不断变化、模棱两可且经常相互矛盾的数据中得出见解时,偶尔会出现不准确的情况。因此,实施适当的评估方法的意义无论怎样强调都不为过,因为它在加强可视化系统的设计和开发方面发挥着举足轻重的作用。本文通过系统地探讨各种评估策略(ES)来评估可视化系统。虽然现有的一些研究已经对一些评估策略进行了研究,但针对可视化研究的全面、系统的综述仍然有限。在这项工作中,我们介绍了七种最先进且广受认可的评估策略,即(1)仪表盘比较;(2)基于洞察力的评估;(3)日志数据分析;(4)李克特量表;(5)定性和定量分析;(6)尼尔森启发式方法;以及(7)眼球跟踪器。此外,本报告还深入探讨了这些 ES 的历史背景,并探讨了这些 ES 的大量应用,揭示了相关的评估实践。通过全面回顾,我们概述并分析了可视化领域的主要评估目标,阐明了它们的演变和内在对比。此外,我们还指出了随着新 ES 的出现而出现的公开挑战,同时还强调了从现有文献中收集到的关键主题,这些主题具有在未来研究中进一步探索的潜力。
{"title":"Strategies for evaluating visual analytics systems: A systematic review and new perspectives","authors":"Md. Rafiqul Islam, Shanjita Akter, Linta Islam, Imran Razzak, Xianzhi Wang, Guandong Xu","doi":"10.1177/14738716231212568","DOIUrl":"https://doi.org/10.1177/14738716231212568","url":null,"abstract":"In recent times, visual analytics systems (VAS) have been used to solve various complex issues in diverse application domains. Nonetheless, an inherent drawback arises from the insufficient evaluation of VAS, resulting in occasional inaccuracies when it comes to analytical reasoning, information synthesis, and deriving insights from vast, ever-changing, ambiguous, and frequently contradictory data. Hence, the significance of implementing an appropriate evaluation methodology cannot be overstated, as it plays a pivotal role in enhancing the design and development of visualization systems. This paper assesses visualization systems by providing a systematic exploration of various evaluation strategies (ES). While several existing studies have examined some ES, the extent of comprehensive and systematic review for visualization research remains limited. In this work, we introduce seven state-of-the-art and widely recognized ES namely (1) dashboard comparison; (2) insight-based evaluation; (3) log data analysis; (4) Likert scales; (5) qualitative and quantitative analysis; (6) Nielsen’s heuristics; and (7) eye trackers. Moreover, it delves into their historical context and explores numerous applications where these ES have been employed, shedding light on the associated evaluation practices. Through our comprehensive review, we overview and analyze the predominant evaluation goals within the visualization community, elucidating their evolution and the inherent contrasts. Additionally, we identify the open challenges that arise with the emergence of new ES, while also highlighting the key themes gleaned from the existing literature that hold potential for further exploration in future studies.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"27 s1","pages":""},"PeriodicalIF":2.3,"publicationDate":"2023-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139150051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-11-24DOI: 10.1177/14738716231213987
Tomás Alves, Ricardo Velhinho, J. Henriques-Calado, Daniel Gonçalves, S. Gama
The anchoring effect is the over-reliance on an initial piece of information when making decisions. It is one of the most pervasive and robust biases. Recently, literature has focused on knowing how influential the anchoring effect is when applied to information visualization, with studies finding its reproducibility in the field. Despite the extensive literature surrounding the anchoring effect’s robustness, there is still a need for research on which individual differences make people more susceptible. We explore how Locus of Control influences visualization’s ubiquitous and resilient anchoring effect. Locus of Control differentiates individuals who believe their life depends on their behavior or actions from those who blame outside factors such as destiny or luck for their life’s outcomes. We focus on the relationship between Locus of Control and the anchoring effect by exposing subjects to an anchor and analyzing their interaction with a complex visualization. Our results show that the anchoring strategies primed individuals and suggest that the Locus of Control plays a role in the susceptibility to the anchoring effect.
锚定效应是指在做决定时过度依赖最初的信息。它是最普遍、最强大的偏见之一。最近,文献重点关注了解锚定效应在应用于信息可视化时的影响力,研究发现它在该领域具有可重复性。尽管有大量文献围绕锚定效应的稳健性展开讨论,但仍有必要研究哪些个体差异使人们更容易受到影响。我们将探讨 "控制感"(Locus of Control)如何影响可视化无处不在且具有弹性的锚定效应。控制感将那些认为自己的人生取决于自己的行为或行动的人与那些将人生结果归咎于命运或运气等外部因素的人区分开来。我们通过让受试者接触锚点并分析他们与复杂可视化的互动,重点研究了控制感与锚定效应之间的关系。我们的研究结果表明,锚定策略对个体产生了引诱作用,并表明控制感在锚定效应的易感性中发挥了作用。
{"title":"Studying the resiliency of the anchoring bias to locus of control in visualization","authors":"Tomás Alves, Ricardo Velhinho, J. Henriques-Calado, Daniel Gonçalves, S. Gama","doi":"10.1177/14738716231213987","DOIUrl":"https://doi.org/10.1177/14738716231213987","url":null,"abstract":"The anchoring effect is the over-reliance on an initial piece of information when making decisions. It is one of the most pervasive and robust biases. Recently, literature has focused on knowing how influential the anchoring effect is when applied to information visualization, with studies finding its reproducibility in the field. Despite the extensive literature surrounding the anchoring effect’s robustness, there is still a need for research on which individual differences make people more susceptible. We explore how Locus of Control influences visualization’s ubiquitous and resilient anchoring effect. Locus of Control differentiates individuals who believe their life depends on their behavior or actions from those who blame outside factors such as destiny or luck for their life’s outcomes. We focus on the relationship between Locus of Control and the anchoring effect by exposing subjects to an anchor and analyzing their interaction with a complex visualization. Our results show that the anchoring strategies primed individuals and suggest that the Locus of Control plays a role in the susceptibility to the anchoring effect.","PeriodicalId":50360,"journal":{"name":"Information Visualization","volume":"32 10","pages":""},"PeriodicalIF":2.3,"publicationDate":"2023-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139240023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}