Pub Date : 2024-06-27DOI: 10.1016/j.visinf.2024.06.001
Christina Stoiber , Daniela Moitzi , Holger Stitz , Florian Grassinger , Anto Silviya Geo Prakash , Dominic Girardi , Marc Streit , Wolfgang Aigner
Visualization onboarding supports users in reading, interpreting, and extracting information from visual data representations. General-purpose onboarding tools and libraries are applicable for explaining a wide range of graphical user interfaces but cannot handle specific visualization requirements. This paper describes a first step towards developing an onboarding library called VisAhoi, which is easy to integrate, extend, semi-automate, reuse, and customize. VisAhoi supports the creation of onboarding elements for different visualization types and datasets. We demonstrate how to extract and describe onboarding instructions using three well-known high-level descriptive visualization grammars — Vega-Lite, Plotly.js, and ECharts. We show the applicability of our library by performing two usage scenarios that describe the integration of VisAhoi into a VA tool for the analysis of high-throughput screening (HTS) data and, second, into a Flourish template to provide an authoring tool for data journalists for a treemap visualization. We provide a supplementary website (https://datavisyn.github.io/visAhoi/) that demonstrates the applicability of VisAhoi to various visualizations, including a bar chart, a horizon graph, a change matrix/heatmap, a scatterplot, and a treemap visualization.
{"title":"VisAhoi: Towards a library to generate and integrate visualization onboarding using high-level visualization grammars","authors":"Christina Stoiber , Daniela Moitzi , Holger Stitz , Florian Grassinger , Anto Silviya Geo Prakash , Dominic Girardi , Marc Streit , Wolfgang Aigner","doi":"10.1016/j.visinf.2024.06.001","DOIUrl":"10.1016/j.visinf.2024.06.001","url":null,"abstract":"<div><p>Visualization onboarding supports users in reading, interpreting, and extracting information from visual data representations. General-purpose onboarding tools and libraries are applicable for explaining a wide range of graphical user interfaces but cannot handle specific visualization requirements. This paper describes a first step towards developing an onboarding library called VisAhoi, which is easy to <em>integrate, extend, semi-automate, reuse, and customize</em>. VisAhoi supports the creation of onboarding elements for different visualization types and datasets. We demonstrate how to extract and describe onboarding instructions using three well-known high-level descriptive visualization grammars — Vega-Lite, Plotly.js, and ECharts. We show the applicability of our library by performing two usage scenarios that describe the integration of VisAhoi into a VA tool for the analysis of high-throughput screening (HTS) data and, second, into a Flourish template to provide an authoring tool for data journalists for a treemap visualization. We provide a supplementary website (<span><span>https://datavisyn.github.io/visAhoi/</span><svg><path></path></svg></span>) that demonstrates the applicability of VisAhoi to various visualizations, including a bar chart, a horizon graph, a change matrix/heatmap, a scatterplot, and a treemap visualization.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 3","pages":"Pages 1-17"},"PeriodicalIF":3.8,"publicationDate":"2024-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000214/pdfft?md5=b500608cf3b6d6a02fdc48334024bff3&pid=1-s2.0-S2468502X24000214-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141954338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.04.003
Yilin Ye , Jianing Hao , Yihan Hou , Zhan Wang , Shishi Xiao , Yuyu Luo , Wei Zeng
Generative AI (GenAI) has witnessed remarkable progress in recent years and demonstrated impressive performance in various generation tasks in different domains such as computer vision and computational design. Many researchers have attempted to integrate GenAI into visualization framework, leveraging the superior generative capacity for different operations. Concurrently, recent major breakthroughs in GenAI like diffusion models and large language models have also drastically increased the potential of GenAI4VIS. From a technical perspective, this paper looks back on previous visualization studies leveraging GenAI and discusses the challenges and opportunities for future research. Specifically, we cover the applications of different types of GenAI methods including sequence, tabular, spatial and graph generation techniques for different tasks of visualization which we summarize into four major stages: data enhancement, visual mapping generation, stylization and interaction. For each specific visualization sub-task, we illustrate the typical data and concrete GenAI algorithms, aiming to provide in-depth understanding of the state-of-the-art GenAI4VIS techniques and their limitations. Furthermore, based on the survey, we discuss three major aspects of challenges and research opportunities including evaluation, dataset, and the gap between end-to-end GenAI methods and visualizations. By summarizing different generation algorithms, their current applications and limitations, this paper endeavors to provide useful insights for future GenAI4VIS research.
{"title":"Generative AI for visualization: State of the art and future directions","authors":"Yilin Ye , Jianing Hao , Yihan Hou , Zhan Wang , Shishi Xiao , Yuyu Luo , Wei Zeng","doi":"10.1016/j.visinf.2024.04.003","DOIUrl":"https://doi.org/10.1016/j.visinf.2024.04.003","url":null,"abstract":"<div><p>Generative AI (GenAI) has witnessed remarkable progress in recent years and demonstrated impressive performance in various generation tasks in different domains such as computer vision and computational design. Many researchers have attempted to integrate GenAI into visualization framework, leveraging the superior generative capacity for different operations. Concurrently, recent major breakthroughs in GenAI like diffusion models and large language models have also drastically increased the potential of GenAI4VIS. From a technical perspective, this paper looks back on previous visualization studies leveraging GenAI and discusses the challenges and opportunities for future research. Specifically, we cover the applications of different types of GenAI methods including sequence, tabular, spatial and graph generation techniques for different tasks of visualization which we summarize into four major stages: data enhancement, visual mapping generation, stylization and interaction. For each specific visualization sub-task, we illustrate the typical data and concrete GenAI algorithms, aiming to provide in-depth understanding of the state-of-the-art GenAI4VIS techniques and their limitations. Furthermore, based on the survey, we discuss three major aspects of challenges and research opportunities including evaluation, dataset, and the gap between end-to-end GenAI methods and visualizations. By summarizing different generation algorithms, their current applications and limitations, this paper endeavors to provide useful insights for future GenAI4VIS research.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 43-66"},"PeriodicalIF":3.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000160/pdfft?md5=c309ceeb991e85de9bb7a69d53c32032&pid=1-s2.0-S2468502X24000160-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141328799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.04.005
Joachim Giesen , Philipp Lucas , Linda Pfeiffer , Laines Schmalwasser , Kai Lawonn
Gaussian mixture models are classical but still popular machine learning models. An appealing feature of Gaussian mixture models is their tractability, that is, they can be learned efficiently and exactly from data, and also support efficient exact inference queries like soft clustering data points. Only seemingly simple, Gaussian mixture models can be hard to understand. There are at least four aspects to understanding Gaussian mixture models, namely, understanding the whole distribution, its individual parts (mixture components), the relationships between the parts, and the interplay of the whole and its parts. In a structured literature review of applications of Gaussian mixture models, we found the need for supporting all four aspects. To identify candidate visualizations that effectively aid the user needs, we structure the available design space along three different representations of Gaussian mixture models, namely as functions, sets of parameters, and sampling processes. From the design space, we implemented three design concepts that visualize the overall distribution together with its components. Finally, we assessed the practical usefulness of the design concepts with respect to the different user needs in expert interviews and an insight-based user study.
{"title":"The whole and its parts: Visualizing Gaussian mixture models","authors":"Joachim Giesen , Philipp Lucas , Linda Pfeiffer , Laines Schmalwasser , Kai Lawonn","doi":"10.1016/j.visinf.2024.04.005","DOIUrl":"10.1016/j.visinf.2024.04.005","url":null,"abstract":"<div><p>Gaussian mixture models are classical but still popular machine learning models. An appealing feature of Gaussian mixture models is their tractability, that is, they can be learned efficiently and exactly from data, and also support efficient exact inference queries like soft clustering data points. Only seemingly simple, Gaussian mixture models can be hard to understand. There are at least four aspects to understanding Gaussian mixture models, namely, understanding the whole distribution, its individual parts (mixture components), the relationships between the parts, and the interplay of the whole and its parts. In a structured literature review of applications of Gaussian mixture models, we found the need for supporting all four aspects. To identify candidate visualizations that effectively aid the user needs, we structure the available design space along three different representations of Gaussian mixture models, namely as functions, sets of parameters, and sampling processes. From the design space, we implemented three design concepts that visualize the overall distribution together with its components. Finally, we assessed the practical usefulness of the design concepts with respect to the different user needs in expert interviews and an insight-based user study.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 67-79"},"PeriodicalIF":3.8,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000196/pdfft?md5=00c8e5cb6796e180f1417fb1e6e4984c&pid=1-s2.0-S2468502X24000196-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141031229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.04.004
Tanja Munz-Körner, Daniel Weiskopf
Dimensionality reduction is often used to project time series data from multidimensional to two-dimensional space to generate visual representations of the temporal evolution. In this context, we address the problem of multidimensional time series visualization by presenting a new method to show and handle projection errors introduced by dimensionality reduction techniques on multidimensional temporal data. For visualization, subsequent time instances are rendered as dots that are connected by lines or curves to indicate the temporal dependencies. However, inevitable projection artifacts may lead to poor visualization quality and misinterpretation of the temporal information. Wrongly projected data points, inaccurate variations in the distances between projected time instances, and intersections of connecting lines could lead to wrong assumptions about the original data. We adapt local and global quality metrics to measure the visual quality along the projected time series, and we introduce a model to assess the projection error at intersecting lines. These serve as a basis for our new uncertainty visualization techniques that use different visual encodings and interactions to indicate, communicate, and work with the visualization uncertainty from projection errors and artifacts along the timeline of data points, their connections, and intersections. Our approach is agnostic to the projection method and works for linear and non-linear dimensionality reduction methods alike.
{"title":"Exploring visual quality of multidimensional time series projections","authors":"Tanja Munz-Körner, Daniel Weiskopf","doi":"10.1016/j.visinf.2024.04.004","DOIUrl":"10.1016/j.visinf.2024.04.004","url":null,"abstract":"<div><p>Dimensionality reduction is often used to project time series data from multidimensional to two-dimensional space to generate visual representations of the temporal evolution. In this context, we address the problem of multidimensional time series visualization by presenting a new method to show and handle projection errors introduced by dimensionality reduction techniques on multidimensional temporal data. For visualization, subsequent time instances are rendered as dots that are connected by lines or curves to indicate the temporal dependencies. However, inevitable projection artifacts may lead to poor visualization quality and misinterpretation of the temporal information. Wrongly projected data points, inaccurate variations in the distances between projected time instances, and intersections of connecting lines could lead to wrong assumptions about the original data. We adapt local and global quality metrics to measure the visual quality along the projected time series, and we introduce a model to assess the projection error at intersecting lines. These serve as a basis for our new uncertainty visualization techniques that use different visual encodings and interactions to indicate, communicate, and work with the visualization uncertainty from projection errors and artifacts along the timeline of data points, their connections, and intersections. Our approach is agnostic to the projection method and works for linear and non-linear dimensionality reduction methods alike.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 27-42"},"PeriodicalIF":3.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000184/pdfft?md5=67711cade8875f71d3b74dad7d012301&pid=1-s2.0-S2468502X24000184-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141142247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.06.002
Jiazhe Wang , Xi Li , Chenlu Li , Di Peng , Arran Zeyu Wang , Yuhui Gu , Xingui Lai , Haifeng Zhang , Xinyue Xu , Xiaoqing Dong , Zhifeng Lin , Jiehui Zhou , Xingyu Liu , Wei Chen
With the incredible growth of the scale and complexity of datasets, creating proper visualizations for users becomes more and more challenging in large datasets. Though several visualization recommendation systems have been proposed, so far, the lack of practical engineering inputs is still a major concern regarding the usage of visualization recommendations in the industry. In this paper, we proposed AVA, an open-sourced web-based framework for Automated Visual Analytics. AVA contains both empiric-driven and insight-driven visualization recommendation methods to meet the demands of creating aesthetic visualizations and understanding expressible insights respectively. The code is available at https://github.com/antvis/AVA.
{"title":"AVA: An automated and AI-driven intelligent visual analytics framework","authors":"Jiazhe Wang , Xi Li , Chenlu Li , Di Peng , Arran Zeyu Wang , Yuhui Gu , Xingui Lai , Haifeng Zhang , Xinyue Xu , Xiaoqing Dong , Zhifeng Lin , Jiehui Zhou , Xingyu Liu , Wei Chen","doi":"10.1016/j.visinf.2024.06.002","DOIUrl":"10.1016/j.visinf.2024.06.002","url":null,"abstract":"<div><p>With the incredible growth of the scale and complexity of datasets, creating proper visualizations for users becomes more and more challenging in large datasets. Though several visualization recommendation systems have been proposed, so far, the lack of practical engineering inputs is still a major concern regarding the usage of visualization recommendations in the industry. In this paper, we proposed <em>AVA</em>, an open-sourced web-based framework for <strong>A</strong>utomated <strong>V</strong>isual <strong>A</strong>nalytics. AVA contains both empiric-driven and insight-driven visualization recommendation methods to meet the demands of creating aesthetic visualizations and understanding expressible insights respectively. The code is available at <span>https://github.com/antvis/AVA</span><svg><path></path></svg>.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 106-114"},"PeriodicalIF":3.8,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000226/pdfft?md5=d535cfeb7d4bca4f8b918b02581ff6a3&pid=1-s2.0-S2468502X24000226-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141410971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.04.006
Jin Xu , Chaojian Zhang , Ming Xie , Xiuxiu Zhan , Luwang Yan , Yubo Tao , Zhigeng Pan
Influence maximization (IM) algorithms play a significant role in hypergraph analysis tasks, such as epidemic control analysis, viral marketing, and social influence analysis, and various IM algorithms have been proposed. The main challenge lies in IM algorithm evaluation, due to the complexity and diversity of the spreading processes of different IM algorithms in different hypergraphs. Existing evaluation methods mainly leverage statistical metrics, such as influence spread, to quantify overall performance, but do not fully unravel spreading characteristics and patterns. In this paper, we propose an exploratory visual analytics system, IMVis, to assist users in exploring and evaluating IM algorithms at the overview, pattern, and node levels. A spreading pattern mining method is first proposed to characterize spreading processes and extract important spreading patterns to facilitate efficient analysis and comparison of IM algorithms. Novel visualization glyphs are designed to comprehensively reveal both temporal and structural features of IM algorithms’ spreading processes in hypergraphs at multiple levels. The effectiveness and usefulness of IMVis are demonstrated through two case studies and expert interviews.
影响最大化(IM)算法在流行病控制分析、病毒营销和社会影响分析等超图分析任务中发挥着重要作用,目前已提出了多种 IM 算法。由于不同 IM 算法在不同超图中的传播过程具有复杂性和多样性,其主要挑战在于 IM 算法的评估。现有的评估方法主要利用影响力传播等统计指标来量化整体性能,但不能完全揭示传播特征和模式。在本文中,我们提出了一个探索性的可视化分析系统--IMVis,以帮助用户从概览、模式和节点三个层面探索和评估 IM 算法。我们首先提出了一种传播模式挖掘方法,以描述传播过程并提取重要的传播模式,从而促进对 IM 算法的有效分析和比较。设计了新颖的可视化字形,以全面揭示 IM 算法在多层次超图中传播过程的时间和结构特征。通过两个案例研究和专家访谈,证明了 IMVis 的有效性和实用性。
{"title":"IMVis: Visual analytics for influence maximization algorithm evaluation in hypergraphs","authors":"Jin Xu , Chaojian Zhang , Ming Xie , Xiuxiu Zhan , Luwang Yan , Yubo Tao , Zhigeng Pan","doi":"10.1016/j.visinf.2024.04.006","DOIUrl":"10.1016/j.visinf.2024.04.006","url":null,"abstract":"<div><p>Influence maximization (IM) algorithms play a significant role in hypergraph analysis tasks, such as epidemic control analysis, viral marketing, and social influence analysis, and various IM algorithms have been proposed. The main challenge lies in IM algorithm evaluation, due to the complexity and diversity of the spreading processes of different IM algorithms in different hypergraphs. Existing evaluation methods mainly leverage statistical metrics, such as influence spread, to quantify overall performance, but do not fully unravel spreading characteristics and patterns. In this paper, we propose an exploratory visual analytics system, IMVis, to assist users in exploring and evaluating IM algorithms at the overview, pattern, and node levels. A spreading pattern mining method is first proposed to characterize spreading processes and extract important spreading patterns to facilitate efficient analysis and comparison of IM algorithms. Novel visualization glyphs are designed to comprehensively reveal both temporal and structural features of IM algorithms’ spreading processes in hypergraphs at multiple levels. The effectiveness and usefulness of IMVis are demonstrated through two case studies and expert interviews.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 13-26"},"PeriodicalIF":3.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000172/pdfft?md5=8a25558f06e02bd13aac06e34e54a160&pid=1-s2.0-S2468502X24000172-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141026551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.05.001
Mengya Zheng, David Lillis, Abraham G. Campbell
Augmented Reality (AR), as a novel data visualization tool, is advantageous in revealing spatial data patterns and data-context associations. Accordingly, recent research has identified AR data visualization as a promising approach to increasing decision-making efficiency and effectiveness. As a result, AR has been applied in various decision support systems to enhance knowledge conveying and comprehension, in which the different data-reality associations have been constructed to aid decision-making.
However, how these AR visualization strategies can enhance different decision support datasets has not been reviewed thoroughly. Especially given the rise of big data in the modern world, this support is critical to decision-making in the coming years. Using AR to embed the decision support data and explanation data into the end user’s physical surroundings and focal contexts avoids isolating the human decision-maker from the relevant data. Integrating the decision-maker’s contexts and the DSS support in AR is a difficult challenge. This paper outlines the current state of the art through a literature review in allowing AR data visualization to support decision-making.
To facilitate the publication classification and analysis, the paper proposes one taxonomy to classify different AR data visualization based on the semantic associations between the AR data and physical context. Based on this taxonomy and a decision support system taxonomy, 37 publications have been classified and analyzed from multiple aspects. One of the contributions of this literature review is a resulting AR visualization taxonomy that can be applied to decision support systems. Along with this novel tool, the paper discusses the current state of the art in this field and indicates possible future challenges and directions that AR data visualization will bring to support decision-making.
增强现实(AR)作为一种新颖的数据可视化工具,在揭示空间数据模式和数据与上下文的关联方面具有优势。因此,最近的研究发现,AR 数据可视化是提高决策效率和效果的一种有前途的方法。因此,AR 已被应用于各种决策支持系统,以加强知识的传达和理解,其中不同的数据-现实关联已被构建以帮助决策。尤其是在大数据兴起的现代社会,这种支持对未来几年的决策至关重要。使用 AR 将决策支持数据和解释数据嵌入最终用户的物理环境和焦点情境中,可避免将人类决策者与相关数据隔离开来。在 AR 中整合决策者的情境和 DSS 支持是一项艰巨的挑战。为了便于对出版物进行分类和分析,本文根据 AR 数据与物理情境之间的语义关联,提出了一种分类法来对不同的 AR 数据可视化进行分类。根据该分类法和决策支持系统分类法,从多个方面对 37 篇出版物进行了分类和分析。本文献综述的贡献之一是提出了可应用于决策支持系统的 AR 可视化分类法。除了这个新颖的工具之外,本文还讨论了该领域的技术现状,并指出了 AR 数据可视化在支持决策方面未来可能面临的挑战和发展方向。
{"title":"Current state of the art and future directions: Augmented reality data visualization to support decision-making","authors":"Mengya Zheng, David Lillis, Abraham G. Campbell","doi":"10.1016/j.visinf.2024.05.001","DOIUrl":"10.1016/j.visinf.2024.05.001","url":null,"abstract":"<div><p>Augmented Reality (AR), as a novel data visualization tool, is advantageous in revealing spatial data patterns and data-context associations. Accordingly, recent research has identified AR data visualization as a promising approach to increasing decision-making efficiency and effectiveness. As a result, AR has been applied in various decision support systems to enhance knowledge conveying and comprehension, in which the different data-reality associations have been constructed to aid decision-making.</p><p>However, how these AR visualization strategies can enhance different decision support datasets has not been reviewed thoroughly. Especially given the rise of big data in the modern world, this support is critical to decision-making in the coming years. Using AR to embed the decision support data and explanation data into the end user’s physical surroundings and focal contexts avoids isolating the human decision-maker from the relevant data. Integrating the decision-maker’s contexts and the DSS support in AR is a difficult challenge. This paper outlines the current state of the art through a literature review in allowing AR data visualization to support decision-making.</p><p>To facilitate the publication classification and analysis, the paper proposes one taxonomy to classify different AR data visualization based on the semantic associations between the AR data and physical context. Based on this taxonomy and a decision support system taxonomy, 37 publications have been classified and analyzed from multiple aspects. One of the contributions of this literature review is a resulting AR visualization taxonomy that can be applied to decision support systems. Along with this novel tool, the paper discusses the current state of the art in this field and indicates possible future challenges and directions that AR data visualization will bring to support decision-making.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 80-105"},"PeriodicalIF":3.8,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000202/pdfft?md5=f80d87851c5113d4a9dd7255cbbe2978&pid=1-s2.0-S2468502X24000202-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141045667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1016/j.visinf.2024.04.002
Yunpeng Chen , Ying Zhao , Xuanjing Li , Jiang Zhang , Jiang Long , Fangfang Zhou
{"title":"Corrigendum to “An open dataset of data lineage graphs for data governance research” [Vis. Inform. 8 (1) (2024) 1-5]","authors":"Yunpeng Chen , Ying Zhao , Xuanjing Li , Jiang Zhang , Jiang Long , Fangfang Zhou","doi":"10.1016/j.visinf.2024.04.002","DOIUrl":"https://doi.org/10.1016/j.visinf.2024.04.002","url":null,"abstract":"","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Page 115"},"PeriodicalIF":3.8,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000159/pdfft?md5=70e77c4a6673309b62e427b282f276e0&pid=1-s2.0-S2468502X24000159-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141486899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-20DOI: 10.1016/j.visinf.2024.04.001
Zhiguang Zhou , Yize Li , Yuna Ni , Weiwen Xu , Guoting Hu , Ying Lai , Peixiong Chen , Weihua Su
Composite index is always derived with the weighted aggregation of hierarchical components, which is widely utilized to distill intricate and multidimensional matters in economic and business statistics. However, the composite indices always present inevitable anomalies at different levels oriented from the calculation and expression processes of hierarchical components, thereby impairing the precise depiction of specific economic issues. In this paper, we propose VisCI, a visualization framework for anomaly detection and interactive optimization of composite index. First, LSTM-AE model is performed to detect anomalies from the lower level to the higher level of the composite index. Then, a comprehensive array of visual cues is designed to visualize anomalies, such as hierarchy and anomaly visualization. In addition, an interactive operation is provided to ensure accurate and efficient index optimization, mitigating the adverse impact of anomalies on index calculation and representation. Finally, we implement a visualization framework with interactive interfaces, facilitating both anomaly detection and intuitive composite index optimization. Case studies based on real-world datasets and expert interviews are conducted to demonstrate the effectiveness of our VisCI in commodity index anomaly exploration and anomaly optimization.
{"title":"VisCI: A visualization framework for anomaly detection and interactive optimization of composite index","authors":"Zhiguang Zhou , Yize Li , Yuna Ni , Weiwen Xu , Guoting Hu , Ying Lai , Peixiong Chen , Weihua Su","doi":"10.1016/j.visinf.2024.04.001","DOIUrl":"10.1016/j.visinf.2024.04.001","url":null,"abstract":"<div><p>Composite index is always derived with the weighted aggregation of hierarchical components, which is widely utilized to distill intricate and multidimensional matters in economic and business statistics. However, the composite indices always present inevitable anomalies at different levels oriented from the calculation and expression processes of hierarchical components, thereby impairing the precise depiction of specific economic issues. In this paper, we propose VisCI, a visualization framework for anomaly detection and interactive optimization of composite index. First, LSTM-AE model is performed to detect anomalies from the lower level to the higher level of the composite index. Then, a comprehensive array of visual cues is designed to visualize anomalies, such as hierarchy and anomaly visualization. In addition, an interactive operation is provided to ensure accurate and efficient index optimization, mitigating the adverse impact of anomalies on index calculation and representation. Finally, we implement a visualization framework with interactive interfaces, facilitating both anomaly detection and intuitive composite index optimization. Case studies based on real-world datasets and expert interviews are conducted to demonstrate the effectiveness of our VisCI in commodity index anomaly exploration and anomaly optimization.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 2","pages":"Pages 1-12"},"PeriodicalIF":3.0,"publicationDate":"2024-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000147/pdfft?md5=e75183915943bf3b9b9ca949c47ab656&pid=1-s2.0-S2468502X24000147-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140793971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Creating realistic materials is essential in the construction of immersive virtual environments. While existing techniques for material capture and conditional generation rely on flash-lit photos, they often produce artifacts when the illumination mismatches the training data. In this study, we introduce DiffMat, a novel diffusion model that integrates the CLIP image encoder and a multi-layer, cross-attention denoising backbone to generate latent materials from images under various illuminations. Using a pre-trained StyleGAN-based material generator, our method converts these latent materials into high-resolution SVBRDF textures, a process that enables a seamless fit into the standard physically based rendering pipeline, reducing the requirements for vast computational resources and expansive datasets. DiffMat surpasses existing generative methods in terms of material quality and variety, and shows adaptability to a broader spectrum of lighting conditions in reference images.
{"title":"DiffMat: Latent diffusion models for image-guided material generation","authors":"Liang Yuan , Dingkun Yan , Suguru Saito , Issei Fujishiro","doi":"10.1016/j.visinf.2023.12.001","DOIUrl":"10.1016/j.visinf.2023.12.001","url":null,"abstract":"<div><p>Creating realistic materials is essential in the construction of immersive virtual environments. While existing techniques for material capture and conditional generation rely on flash-lit photos, they often produce artifacts when the illumination mismatches the training data. In this study, we introduce DiffMat, a novel diffusion model that integrates the CLIP image encoder and a multi-layer, cross-attention denoising backbone to generate latent materials from images under various illuminations. Using a pre-trained StyleGAN-based material generator, our method converts these latent materials into high-resolution SVBRDF textures, a process that enables a seamless fit into the standard physically based rendering pipeline, reducing the requirements for vast computational resources and expansive datasets. DiffMat surpasses existing generative methods in terms of material quality and variety, and shows adaptability to a broader spectrum of lighting conditions in reference images.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 1","pages":"Pages 6-14"},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000019/pdfft?md5=fb0200304a9b292debbf18a3162d10e8&pid=1-s2.0-S2468502X24000019-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139396034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}