首页 > 最新文献

Visual Informatics最新文献

英文 中文
Key-isovalue selection and hierarchical exploration visualization of weather forecast ensembles 天气预报集合的键等值选择与分层探索可视化
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-03-01 DOI: 10.1016/j.visinf.2025.02.001
Feng Zhou, Hao Hu, Fengjie Wang, Jiamin Zhu, Wenwen Gao, Min Zhu
Weather forecast ensembles are commonly used to assess the uncertainty and confidence of weather predictions. Conventional methods in meteorology often employ ensemble mean and standard deviation plots, as well as spaghetti plots, to visualize ensemble data. However, these methods suffer from significant information loss and visual clutter. In this paper, we propose a new approach for uncertainty visualization of weather forecast ensembles, including isovalue selection based on information loss and hierarchical visualization that integrates visual abstraction and detail preservation. Our approach uses non-uniform downsampling to select key-isovalues and provides an interactive visualization method based on hierarchical clustering. Firstly, we sample key-isovalues by contour probability similarity and determine the optimal sampling number using an information loss curve. Then, the corresponding isocontours are presented to guide users in selecting key-isovalues. Once the isovalue is chosen, we perform agglomerative hierarchical clustering on the isocontours based on signed distance fields and generate visual abstractions for each isocontour cluster to avoid visual clutter. We link a bubble tree to the visual abstractions to explore the details of isocontour clusters at different levels. We demonstrate the utility of our approach through two case studies with meteorological experts on real-world data. We further validate its effectiveness by quantitatively assessing information loss and visual clutter. Additionally, we confirm its usability through expert evaluation.
天气预报集合通常用于评估天气预报的不确定性和可信度。气象学的传统方法通常采用集合均值和标准偏差图以及意大利面图来可视化集合数据。然而,这些方法存在严重的信息丢失和视觉混乱。在本文中,我们提出了一种新的天气预报集合的不确定性可视化方法,包括基于信息损失的等值选择和集成了视觉抽象和细节保留的分层可视化。我们的方法使用非均匀下采样来选择键等值,并提供了一种基于分层聚类的交互式可视化方法。首先,利用轮廓概率相似度对键等值进行采样,利用信息损失曲线确定最优采样次数;然后给出相应的等等值线,指导用户选择关键等等值线。选取等高值后,基于符号距离域对等高线进行聚类,并对每个等高线聚类生成视觉抽象,避免视觉混乱。我们将气泡树与视觉抽象联系起来,以探索不同层次的等轮廓簇的细节。我们通过气象专家对真实世界数据的两个案例研究,展示了我们方法的实用性。我们通过定量评估信息丢失和视觉混乱来进一步验证其有效性。此外,我们通过专家评估确认其可用性。
{"title":"Key-isovalue selection and hierarchical exploration visualization of weather forecast ensembles","authors":"Feng Zhou,&nbsp;Hao Hu,&nbsp;Fengjie Wang,&nbsp;Jiamin Zhu,&nbsp;Wenwen Gao,&nbsp;Min Zhu","doi":"10.1016/j.visinf.2025.02.001","DOIUrl":"10.1016/j.visinf.2025.02.001","url":null,"abstract":"<div><div>Weather forecast ensembles are commonly used to assess the uncertainty and confidence of weather predictions. Conventional methods in meteorology often employ ensemble mean and standard deviation plots, as well as spaghetti plots, to visualize ensemble data. However, these methods suffer from significant information loss and visual clutter. In this paper, we propose a new approach for uncertainty visualization of weather forecast ensembles, including isovalue selection based on information loss and hierarchical visualization that integrates visual abstraction and detail preservation. Our approach uses non-uniform downsampling to select key-isovalues and provides an interactive visualization method based on hierarchical clustering. Firstly, we sample key-isovalues by contour probability similarity and determine the optimal sampling number using an information loss curve. Then, the corresponding isocontours are presented to guide users in selecting key-isovalues. Once the isovalue is chosen, we perform agglomerative hierarchical clustering on the isocontours based on signed distance fields and generate visual abstractions for each isocontour cluster to avoid visual clutter. We link a bubble tree to the visual abstractions to explore the details of isocontour clusters at different levels. We demonstrate the utility of our approach through two case studies with meteorological experts on real-world data. We further validate its effectiveness by quantitatively assessing information loss and visual clutter. Additionally, we confirm its usability through expert evaluation.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 58-70"},"PeriodicalIF":3.8,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What about thematic information? An analysis of the multidimensional visualization of individual mobility 主题信息呢?个人流动的多维可视化分析
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-03-01 DOI: 10.1016/j.visinf.2025.02.002
Aline Menin , Clément Quere , Jorge Wagner , Sonia Chardonnel , Paule-Annick Davoine , Wolfgang Stuerzlinger , Carla Maria Dal Sasso Freitas , Luciana Nedel , Marco Winckler
This paper reviews the literature on the visualization of individual mobility data, with a focus on thematic integration. It emphasizes the importance of visualization in understanding mobility patterns within a population and how it helps mobility experts address domain-specific questions. We analyze 38 papers published between 2010 and 2024 in GIS and VIS venues that describe visualizations of multidimensional data related to individual movements in urban environments, concentrating on individual mobility rather than traffic data. Our primary aim is to report advances in interactive visualization for individual mobility analysis, particularly regarding the representation of thematic information about people’s motivations for mobility. Our findings indicate that the thematic dimension is only partially represented in the literature, despite its critical significance in transportation. This gap often stems from the challenge of identifying data sources that inherently provide this information, necessitating visualization designers and developers to navigate multiple, heterogeneous data sources. We identify the strengths and limitations of existing visualizations and suggest potential research directions for the field.
本文综述了关于个人移动数据可视化的文献,重点介绍了主题集成。它强调了可视化在理解人口流动模式中的重要性,以及它如何帮助流动专家解决特定领域的问题。我们分析了2010年至2024年间发表在GIS和VIS领域的38篇论文,这些论文描述了与城市环境中个人运动相关的多维数据的可视化,重点关注个人移动性而不是交通数据。我们的主要目的是报告个人流动性分析的交互式可视化方面的进展,特别是关于人们流动性动机的主题信息的表示。我们的研究结果表明,尽管主题维度在交通运输中具有重要意义,但它在文献中仅得到部分体现。这种差距通常源于识别本质上提供此信息的数据源的挑战,这要求可视化设计人员和开发人员导航多个异构数据源。我们确定现有可视化的优势和局限性,并提出该领域潜在的研究方向。
{"title":"What about thematic information? An analysis of the multidimensional visualization of individual mobility","authors":"Aline Menin ,&nbsp;Clément Quere ,&nbsp;Jorge Wagner ,&nbsp;Sonia Chardonnel ,&nbsp;Paule-Annick Davoine ,&nbsp;Wolfgang Stuerzlinger ,&nbsp;Carla Maria Dal Sasso Freitas ,&nbsp;Luciana Nedel ,&nbsp;Marco Winckler","doi":"10.1016/j.visinf.2025.02.002","DOIUrl":"10.1016/j.visinf.2025.02.002","url":null,"abstract":"<div><div>This paper reviews the literature on the visualization of individual mobility data, with a focus on thematic integration. It emphasizes the importance of visualization in understanding mobility patterns within a population and how it helps mobility experts address domain-specific questions. We analyze 38 papers published between 2010 and 2024 in GIS and VIS venues that describe visualizations of multidimensional data related to individual movements in urban environments, concentrating on individual mobility rather than traffic data. Our primary aim is to report advances in interactive visualization for individual mobility analysis, particularly regarding the representation of thematic information about people’s motivations for mobility. Our findings indicate that the thematic dimension is only partially represented in the literature, despite its critical significance in transportation. This gap often stems from the challenge of identifying data sources that inherently provide this information, necessitating visualization designers and developers to navigate multiple, heterogeneous data sources. We identify the strengths and limitations of existing visualizations and suggest potential research directions for the field.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 99-115"},"PeriodicalIF":3.8,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ClayVolume: A progressive refinement interaction system for immersive visualization ClayVolume:一个用于沉浸式可视化的渐进细化交互系统
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-03-01 DOI: 10.1016/j.visinf.2025.01.003
Zhenyuan Wang , Qing Zhao , Yue Zhang , Jinhui Zhang , Guihua Shan , Xiao Zhou , Dong Tian
Immersive visualization has become an important tool for discovering hidden patterns and obtaining insights from data. Target acquisition in immersive visualization is a fundamental step in visual analysis. However, limited visual encoding attributes and the presence of stacking and occlusion in immersive environments pose challenges in discovering valuable targets and making unambiguous selections. In this paper, we present ClayVolume, an interactive system designed for immersive visualization. It comprises metaphorical tools for customizing regions of interest (ROIs) and multiple views that serve as interactive and analytical mediums. ClayVolume empowers analysts to efficiently acquire valuable targets through a progressive refinement of interactive methods, enabling further extraction of insights. We evaluate ClayVolume in the scenario of immersive visualization of network data and perform a comparative analysis of its performance against other techniques in target selection tasks. The results indicate that ClayVolume enables flexible target selection in immersive visualization and provides fast target discovery and localization capabilities.
沉浸式可视化已经成为发现隐藏模式和从数据中获得洞察力的重要工具。沉浸式可视化中的目标获取是可视化分析的基本步骤。然而,在沉浸式环境中,有限的视觉编码属性和叠加和遮挡的存在给发现有价值的目标和做出明确的选择带来了挑战。在本文中,我们提出了ClayVolume,一个为沉浸式可视化设计的交互系统。它包括用于定制感兴趣区域(roi)的隐喻工具和作为交互和分析媒介的多个视图。ClayVolume使分析人员能够通过交互式方法的逐步改进有效地获取有价值的目标,从而进一步提取见解。我们在网络数据沉浸式可视化场景中评估了ClayVolume,并对其在目标选择任务中的性能与其他技术进行了比较分析。结果表明,ClayVolume能够在沉浸式可视化中实现灵活的目标选择,并提供快速的目标发现和定位能力。
{"title":"ClayVolume: A progressive refinement interaction system for immersive visualization","authors":"Zhenyuan Wang ,&nbsp;Qing Zhao ,&nbsp;Yue Zhang ,&nbsp;Jinhui Zhang ,&nbsp;Guihua Shan ,&nbsp;Xiao Zhou ,&nbsp;Dong Tian","doi":"10.1016/j.visinf.2025.01.003","DOIUrl":"10.1016/j.visinf.2025.01.003","url":null,"abstract":"<div><div>Immersive visualization has become an important tool for discovering hidden patterns and obtaining insights from data. Target acquisition in immersive visualization is a fundamental step in visual analysis. However, limited visual encoding attributes and the presence of stacking and occlusion in immersive environments pose challenges in discovering valuable targets and making unambiguous selections. In this paper, we present ClayVolume, an interactive system designed for immersive visualization. It comprises metaphorical tools for customizing regions of interest (ROIs) and multiple views that serve as interactive and analytical mediums. ClayVolume empowers analysts to efficiently acquire valuable targets through a progressive refinement of interactive methods, enabling further extraction of insights. We evaluate ClayVolume in the scenario of immersive visualization of network data and perform a comparative analysis of its performance against other techniques in target selection tasks. The results indicate that ClayVolume enables flexible target selection in immersive visualization and provides fast target discovery and localization capabilities.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 71-83"},"PeriodicalIF":3.8,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EmotionLens: Interactive visual exploration of the circumplex emotion space in literary works via affective word clouds EmotionLens:通过情感词云对文学作品中复杂的情感空间进行交互式视觉探索
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-03-01 DOI: 10.1016/j.visinf.2025.02.003
Bingyuan Wang , Qing Shi , Xiaohan Wang , You Zhou , Wei Zeng , Zeyu Wang
Emotion (e.g., valence and arousal) is an important factor in literature (e.g., poetry and prose), and has rich values for plotting the life and knowledge of historical figures and appreciating the aesthetics of literary works. Currently, digital humanities and computational literature apply data statistics extensively in emotion analysis but lack visual analytics for efficient exploration. To fill the gap, we propose a user-centric approach that integrates advanced machine learning models and intuitive visualization for emotion analysis in literature. We make three main contributions. First, we consolidate a new emotion dataset of literary works in different periods, literary genres, and language contexts, augmented with fine-grained valence and arousal labels. Next, we design an interactive visual analytic system named EmotionLens, which allows users to perform multi-granularity (e.g., individual, group, society) and multi-faceted (e.g., distribution, chronology, correlation) analyses of literary emotions, supporting both exploratory and confirmatory approaches in digital humanities. Specifically, we introduce a novel affective word cloud with augmented word weight, position, and color, to facilitate literary text analysis from an emotional perspective. To validate the usability and effectiveness of EmotionLens, we provide two consecutive case studies, two user studies, and interviews with experts from different domains. Our results show that EmotionLens bridges literary text, emotion, and various other attributes, enables efficient knowledge discovery in massive data, and facilitates raising and validating domain-specific hypotheses in literature.
情感(如价、情)是文学(如诗、文)的重要因素,在描绘历史人物的生活、知识和欣赏文学作品的美学方面具有丰富的价值。目前,数字人文学科和计算文学将数据统计广泛应用于情感分析,但缺乏可视化分析来进行有效的探索。为了填补这一空白,我们提出了一种以用户为中心的方法,该方法集成了先进的机器学习模型和直观的可视化,用于文献中的情感分析。我们做出了三个主要贡献。首先,我们整合了一个新的情感数据集,该数据集包含不同时期、不同文学类型和不同语言背景的文学作品,并添加了细粒度的价态和唤醒标签。接下来,我们设计了一个名为EmotionLens的交互式视觉分析系统,该系统允许用户对文学情感进行多粒度(例如,个人,群体,社会)和多方面(例如,分布,时间顺序,相关性)分析,支持数字人文学科的探索性和验证性方法。具体来说,我们引入了一种新的情感词云,增强了词的权重、位置和颜色,以促进从情感角度分析文学文本。为了验证EmotionLens的可用性和有效性,我们提供了两个连续的案例研究,两个用户研究,并采访了来自不同领域的专家。我们的研究结果表明,EmotionLens在文学文本、情感和各种其他属性之间架起了桥梁,能够在海量数据中高效地发现知识,并有助于提出和验证文学中特定领域的假设。
{"title":"EmotionLens: Interactive visual exploration of the circumplex emotion space in literary works via affective word clouds","authors":"Bingyuan Wang ,&nbsp;Qing Shi ,&nbsp;Xiaohan Wang ,&nbsp;You Zhou ,&nbsp;Wei Zeng ,&nbsp;Zeyu Wang","doi":"10.1016/j.visinf.2025.02.003","DOIUrl":"10.1016/j.visinf.2025.02.003","url":null,"abstract":"<div><div>Emotion (e.g., valence and arousal) is an important factor in literature (e.g., poetry and prose), and has rich values for plotting the life and knowledge of historical figures and appreciating the aesthetics of literary works. Currently, digital humanities and computational literature apply data statistics extensively in emotion analysis but lack visual analytics for efficient exploration. To fill the gap, we propose a user-centric approach that integrates advanced machine learning models and intuitive visualization for emotion analysis in literature. We make three main contributions. First, we consolidate a new emotion dataset of literary works in different periods, literary genres, and language contexts, augmented with fine-grained valence and arousal labels. Next, we design an interactive visual analytic system named <em>EmotionLens</em>, which allows users to perform multi-granularity (e.g., individual, group, society) and multi-faceted (e.g., distribution, chronology, correlation) analyses of literary emotions, supporting both exploratory and confirmatory approaches in digital humanities. Specifically, we introduce a novel affective word cloud with augmented word weight, position, and color, to facilitate literary text analysis from an emotional perspective. To validate the usability and effectiveness of <em>EmotionLens</em>, we provide two consecutive case studies, two user studies, and interviews with experts from different domains. Our results show that <em>EmotionLens</em> bridges literary text, emotion, and various other attributes, enables efficient knowledge discovery in massive data, and facilitates raising and validating domain-specific hypotheses in literature.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 84-98"},"PeriodicalIF":3.8,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging personality as a proxy of perceived transparency in hierarchical visualizations 在层次可视化中,利用个性作为感知透明度的代理
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-02-22 DOI: 10.1016/j.visinf.2025.01.002
Tomás Alves , Carlota Dias , Daniel Gonçalves , Sandra Gama
Understanding which factors affect information visualization transparency continues to be one of the most relevant challenges in current research, especially since trust models how users build on the knowledge and use it. This work extends the current body of research by studying the user’s subjective evaluation of the visualization transparency of hierarchical charts through the clarity, coverage, and look and feel dimensions. Additionally, we extend the user profile to better understand whether personality facets manifest a biasing effect on the trust-building process. Our results show that the data encodings do not affect how users perceive visualization transparency while controlling for personality factors. Regarding personality, the propensity to trust affects how they judge the clarity of a hierarchical chart. Our findings provide new insights into the research challenges of measuring trust and understanding the transparency of information visualization. Specifically, we explore how personality factors manifest in this trust-building relationship and user interaction within visualization systems.
了解哪些因素会影响信息可视化透明度仍然是当前研究中最相关的挑战之一,特别是因为信任建模了用户如何建立和使用知识。这项工作通过研究用户通过清晰度、覆盖范围和外观和感觉维度对分层图表可视化透明度的主观评价,扩展了当前的研究主体。此外,我们扩展了用户档案,以更好地了解人格方面是否在信任建立过程中表现出偏见效应。我们的研究结果表明,在控制个性因素的情况下,数据编码不会影响用户对可视化透明度的感知。在人格方面,信任倾向会影响他们对等级图表清晰度的判断。我们的研究结果为衡量信任和理解信息可视化透明度的研究挑战提供了新的见解。具体来说,我们探讨了人格因素如何在可视化系统中表现出这种信任建立关系和用户交互。
{"title":"Leveraging personality as a proxy of perceived transparency in hierarchical visualizations","authors":"Tomás Alves ,&nbsp;Carlota Dias ,&nbsp;Daniel Gonçalves ,&nbsp;Sandra Gama","doi":"10.1016/j.visinf.2025.01.002","DOIUrl":"10.1016/j.visinf.2025.01.002","url":null,"abstract":"<div><div>Understanding which factors affect information visualization transparency continues to be one of the most relevant challenges in current research, especially since trust models how users build on the knowledge and use it. This work extends the current body of research by studying the user’s subjective evaluation of the visualization transparency of hierarchical charts through the clarity, coverage, and look and feel dimensions. Additionally, we extend the user profile to better understand whether personality facets manifest a biasing effect on the trust-building process. Our results show that the data encodings do not affect how users perceive visualization transparency while controlling for personality factors. Regarding personality, the propensity to trust affects how they judge the clarity of a hierarchical chart. Our findings provide new insights into the research challenges of measuring trust and understanding the transparency of information visualization. Specifically, we explore how personality factors manifest in this trust-building relationship and user interaction within visualization systems.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 43-57"},"PeriodicalIF":3.8,"publicationDate":"2025-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143464086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visual comparative analytics of multimodal transportation 多式联运的视觉比较分析
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-01-16 DOI: 10.1016/j.visinf.2025.01.001
Zikun Deng , Haoming Chen , Qing-Long Lu , Zicheng Su , Tobias Schreck , Jie Bao , Yi Cai
Contemporary urban transportation systems frequently depend on a variety of modes to provide residents with travel services. Understanding a multimodal transportation system is pivotal for devising well-informed planning; however, it is also inherently challenging for traffic analysts and planners. This challenge stems from the necessity of evaluating and contrasting the quality of transportation services across multiple modes. Existing methods are constrained in offering comprehensive insights into the system, primarily due to the inadequacy of multimodal traffic data necessary for fair comparisons and their inability to equip analysts and planners with the means for exploration and reasoned analysis within the urban spatial context. To this end, we first acquire sufficient multimodal trips leveraging well-established navigation platforms that can estimate the routes with the least travel time given an origin and a destination (an OD pair). We also propose TraDyssey, a visual analytics system that enables analysts and planners to evaluate and compare multiple modes by exploring acquired massive multimodal trips. TraDyssey follows a streamlined query-and-explore workflow supported by user-friendly and effective interactive visualizations. Specifically, a revisited difference-aware parallel coordinate plot (PCP) is designed for overall mode comparisons based on multimodal trips. Trip groups can be flexibly queried on the PCP based on differential features across modes. The queried trips are then organized and presented on a geographic map by OD pairs, forming a group-OD-trip hierarchy of visual exploration. Domain experts gained valuable insights into transportation planning through real-world case studies using TraDyssey.
当代城市交通系统通常依靠多种模式为居民提供出行服务。了解多式联运系统对于制定明智的规划至关重要,但对交通分析师和规划师来说,这本身也是一项挑战。这一挑战源于对多种交通方式的交通服务质量进行评估和对比的必要性。现有方法在提供对系统的全面见解方面受到限制,主要原因是缺乏进行公平比较所需的多模式交通数据,以及无法为分析师和规划师提供在城市空间背景下进行探索和合理分析的手段。为此,我们首先利用成熟的导航平台获取足够的多式联运出行数据,这些平台可以根据起点和终点(OD 对)估算出旅行时间最少的路线。我们还提出了 TraDyssey,这是一个可视化分析系统,使分析师和规划师能够通过探索获取的大量多式联运行程来评估和比较多种模式。TraDyssey 采用简化的查询和探索工作流程,并辅以用户友好和有效的交互式可视化。具体来说,基于多式联运的整体模式比较设计了一个重新设计的差异感知平行坐标图(PCP)。根据不同模式的差异特征,可以在平行坐标图上灵活地查询行程组。然后,查询到的行程按 OD 对在地理地图上进行组织和展示,形成一个可视化探索的组-OD-行程层次结构。领域专家通过使用 TraDyssey 进行实际案例研究,对交通规划获得了宝贵的见解。
{"title":"Visual comparative analytics of multimodal transportation","authors":"Zikun Deng ,&nbsp;Haoming Chen ,&nbsp;Qing-Long Lu ,&nbsp;Zicheng Su ,&nbsp;Tobias Schreck ,&nbsp;Jie Bao ,&nbsp;Yi Cai","doi":"10.1016/j.visinf.2025.01.001","DOIUrl":"10.1016/j.visinf.2025.01.001","url":null,"abstract":"<div><div>Contemporary urban transportation systems frequently depend on a variety of modes to provide residents with travel services. Understanding a multimodal transportation system is pivotal for devising well-informed planning; however, it is also inherently challenging for traffic analysts and planners. This challenge stems from the necessity of evaluating and contrasting the quality of transportation services across multiple modes. Existing methods are constrained in offering comprehensive insights into the system, primarily due to the inadequacy of multimodal traffic data necessary for fair comparisons and their inability to equip analysts and planners with the means for exploration and reasoned analysis within the urban spatial context. To this end, we first acquire sufficient multimodal trips leveraging well-established navigation platforms that can estimate the routes with the least travel time given an origin and a destination (an OD pair). We also propose TraDyssey, a visual analytics system that enables analysts and planners to evaluate and compare multiple modes by exploring acquired massive multimodal trips. TraDyssey follows a streamlined query-and-explore workflow supported by user-friendly and effective interactive visualizations. Specifically, a revisited difference-aware parallel coordinate plot (PCP) is designed for overall mode comparisons based on multimodal trips. Trip groups can be flexibly queried on the PCP based on differential features across modes. The queried trips are then organized and presented on a geographic map by OD pairs, forming a group-OD-trip hierarchy of visual exploration. Domain experts gained valuable insights into transportation planning through real-world case studies using TraDyssey.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 18-30"},"PeriodicalIF":3.8,"publicationDate":"2025-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143445454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Out-of-focus artifacts mitigation and autofocus methods for 3D displays 3D显示器的失焦伪影缓解和自动对焦方法
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-20 DOI: 10.1016/j.visinf.2024.12.001
T. Chlubna , T. Milet , P. Zemčík
This paper proposes a novel content-aware method for automatic focusing of the scene on a 3D display. The method addresses a common problem that visualized content is often out of focus, which adversely affects perceived 3D content. The method outperforms existing focusing method, having the error lower by almost 30%. The existing and novel focusing is extended with depth-of-field enhancement of the scene to mitigate out-of-focus artifacts. The relation between the total depth range of the scene and the visual quality of the result is discussed and evaluated according to human perception experiments. A space-warping method for synthetic scenes is proposed to reduce out-of-focus artifacts while maintaining the scene appearance. A user study was conducted to evaluate the proposed methods and identify the crucial parameters in the scene-focusing process on the 3D stereoscopic display by Looking Glass Factory. The study confirmed the efficiency of the proposals and discovered that the depth-of-field artifact mitigation might not be suitable for all scenes despite theoretical hypotheses. The overall proposal of this paper is a set of methods that can be used to produce the best user experience with an arbitrary scene displayed on a 3D display.
提出了一种基于内容感知的三维场景自动对焦方法。该方法解决了可视化内容经常失焦的常见问题,这对感知的3D内容产生了不利影响。该方法优于现有的调焦方法,误差降低了近30%。现有的和新的聚焦扩展与景深增强的场景,以减轻失焦的伪影。根据人的感知实验,讨论了场景的总深度范围与结果视觉质量之间的关系,并对其进行了评价。提出了一种用于合成场景的空间扭曲方法,在保持场景外观的同时减少失焦伪影。通过用户研究,对所提出的方法进行了评估,并确定了Looking Glass Factory 3D立体显示器场景聚焦过程中的关键参数。研究证实了这些建议的有效性,并发现尽管有理论假设,但景深伪影缓解可能并不适用于所有场景。本文的总体建议是一组方法,可用于在3D显示器上显示任意场景以产生最佳用户体验。
{"title":"Out-of-focus artifacts mitigation and autofocus methods for 3D displays","authors":"T. Chlubna ,&nbsp;T. Milet ,&nbsp;P. Zemčík","doi":"10.1016/j.visinf.2024.12.001","DOIUrl":"10.1016/j.visinf.2024.12.001","url":null,"abstract":"<div><div>This paper proposes a novel content-aware method for automatic focusing of the scene on a 3D display. The method addresses a common problem that visualized content is often out of focus, which adversely affects perceived 3D content. The method outperforms existing focusing method, having the error lower by almost 30%. The existing and novel focusing is extended with depth-of-field enhancement of the scene to mitigate out-of-focus artifacts. The relation between the total depth range of the scene and the visual quality of the result is discussed and evaluated according to human perception experiments. A space-warping method for synthetic scenes is proposed to reduce out-of-focus artifacts while maintaining the scene appearance. A user study was conducted to evaluate the proposed methods and identify the crucial parameters in the scene-focusing process on the 3D stereoscopic display by Looking Glass Factory. The study confirmed the efficiency of the proposals and discovered that the depth-of-field artifact mitigation might not be suitable for all scenes despite theoretical hypotheses. The overall proposal of this paper is a set of methods that can be used to produce the best user experience with an arbitrary scene displayed on a 3D display.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 31-42"},"PeriodicalIF":3.8,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143445455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transforming cinematography lighting education in the metaverse 在虚拟世界中改变电影照明教育
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-05 DOI: 10.1016/j.visinf.2024.11.003
Xian Xu , Wai Tong , Zheng Wei , Meng Xia , Lik-Hang Lee , Huamin Qu
Lighting education is a foundational component of cinematography education. However, many art schools do not have expensive soundstages for traditional cinematography lessons. Migrating physical setups to virtual experiences is a potential solution driven by metaverse initiatives. Yet there is still a lack of knowledge on the design of a VR system for teaching cinematography. We first analyzed the educational needs for cinematography lighting education by conducting interviews with six cinematography professionals from academia and industry. Accordingly, we presented Art Mirror, a VR soundstage for teachers and students to emulate cinematography lighting in virtual scenarios. We evaluated Art Mirror from the aspects of usability, realism, presence, sense of agency, and collaboration. Sixteen participants were invited to take a cinematography lighting course and assess the design elements of Art Mirror. Our results demonstrate that Art Mirror is usable and useful for cinematography lighting education, which sheds light on the design of VR cinematography education.
灯光教育是电影教育的基础组成部分。然而,许多艺术学校没有昂贵的传统电影摄影课程的摄影棚。将物理设置迁移到虚拟体验是一种潜在的解决方案,由虚拟活动驱动。然而,关于VR系统在电影摄影教学中的设计仍然缺乏知识。我们首先透过访谈六位来自学术界和业界的电影摄影专业人士,分析电影摄影灯光教育的教育需求。因此,我们提出了Art Mirror,这是一个VR声场,供教师和学生在虚拟场景中模拟电影摄影灯光。我们从可用性、现实性、存在感、代理感和协作性等方面对Art Mirror进行了评估。16名参与者被邀请参加电影摄影灯光课程,并评估艺术镜子的设计元素。我们的研究结果表明,艺术镜子在电影灯光教育中是可用的和有用的,这对VR电影灯光教育的设计有一定的启示。
{"title":"Transforming cinematography lighting education in the metaverse","authors":"Xian Xu ,&nbsp;Wai Tong ,&nbsp;Zheng Wei ,&nbsp;Meng Xia ,&nbsp;Lik-Hang Lee ,&nbsp;Huamin Qu","doi":"10.1016/j.visinf.2024.11.003","DOIUrl":"10.1016/j.visinf.2024.11.003","url":null,"abstract":"<div><div>Lighting education is a foundational component of cinematography education. However, many art schools do not have expensive soundstages for traditional cinematography lessons. Migrating physical setups to virtual experiences is a potential solution driven by metaverse initiatives. Yet there is still a lack of knowledge on the design of a VR system for teaching cinematography. We first analyzed the educational needs for cinematography lighting education by conducting interviews with six cinematography professionals from academia and industry. Accordingly, we presented <em>Art Mirror</em>, a VR soundstage for teachers and students to emulate cinematography lighting in virtual scenarios. We evaluated <em>Art Mirror</em> from the aspects of usability, realism, presence, sense of agency, and collaboration. Sixteen participants were invited to take a cinematography lighting course and assess the design elements of <em>Art Mirror</em>. Our results demonstrate that <em>Art Mirror</em> is usable and useful for cinematography lighting education, which sheds light on the design of VR cinematography education.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"9 1","pages":"Pages 1-17"},"PeriodicalIF":3.8,"publicationDate":"2024-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143437611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ArtEyer: Enriching GPT-based agents with contextual data visualizations for fine art authentication ArtEyer:为美术认证丰富基于gpt的代理与上下文数据可视化
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-01 DOI: 10.1016/j.visinf.2024.11.001
Tan Tang , Yanhong Wu , Junming Gao , Kejia Ruan , Yanjie Zhang , Shuainan Ye , Yingcai Wu , Xiaojiao Chen
Fine art authentication plays a significant role in protecting cultural heritage and ensuring the integrity of artworks. Traditional authentication methods require professionals to collect many reference materials and conduct detailed analyses. To ease the difficulty, we collaborate with domain experts to develop a GPT-based agent, namely ArtEyer, that offers accurate attributions, determines the origin and authorship, and executes visual analytics. Despite the convenience of the conversational user interface, novice users may still face challenges due to the hallucination issue and the steep learning curve associated with prompting. To face these obstacles, we propose a novel solution that places interactive data visualizations into the conversations. We create contextual visualizations from an external domain-dependent database to ensure data trustworthiness and allow users to provide precise instructions to the agent by interacting directly with these visualizations, thus overcoming the vagueness inherent in natural language-based prompting. We evaluate ArtEyer through an in-lab user study and demonstrate its usage with a real-world case.
美术鉴定在保护文化遗产、保证艺术品完整性方面具有重要作用。传统的鉴定方法需要专业人员收集大量的参考资料并进行详细的分析。为了减轻困难,我们与领域专家合作开发了一个基于gpt的代理,即ArtEyer,它提供准确的归属,确定来源和作者,并执行可视化分析。尽管会话式用户界面很方便,但由于幻觉问题和与提示相关的陡峭学习曲线,新手用户可能仍然面临挑战。为了面对这些障碍,我们提出了一种新颖的解决方案,将交互式数据可视化放入对话中。我们从外部领域相关数据库创建上下文可视化,以确保数据的可信度,并允许用户通过直接与这些可视化交互向代理提供精确的指令,从而克服基于自然语言的提示固有的模糊性。我们通过实验室用户研究来评估ArtEyer,并通过实际案例展示其使用情况。
{"title":"ArtEyer: Enriching GPT-based agents with contextual data visualizations for fine art authentication","authors":"Tan Tang ,&nbsp;Yanhong Wu ,&nbsp;Junming Gao ,&nbsp;Kejia Ruan ,&nbsp;Yanjie Zhang ,&nbsp;Shuainan Ye ,&nbsp;Yingcai Wu ,&nbsp;Xiaojiao Chen","doi":"10.1016/j.visinf.2024.11.001","DOIUrl":"10.1016/j.visinf.2024.11.001","url":null,"abstract":"<div><div>Fine art authentication plays a significant role in protecting cultural heritage and ensuring the integrity of artworks. Traditional authentication methods require professionals to collect many reference materials and conduct detailed analyses. To ease the difficulty, we collaborate with domain experts to develop a GPT-based agent, namely ArtEyer, that offers accurate attributions, determines the origin and authorship, and executes visual analytics. Despite the convenience of the conversational user interface, novice users may still face challenges due to the hallucination issue and the steep learning curve associated with prompting. To face these obstacles, we propose a novel solution that places interactive data visualizations into the conversations. We create contextual visualizations from an external domain-dependent database to ensure data trustworthiness and allow users to provide precise instructions to the agent by interacting directly with these visualizations, thus overcoming the vagueness inherent in natural language-based prompting. We evaluate ArtEyer through an in-lab user study and demonstrate its usage with a real-world case.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 4","pages":"Pages 48-59"},"PeriodicalIF":3.8,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143098848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Computer Vision in Augmented, Virtual, Mixed and Extended Reality environments—A bibliometric review 增强、虚拟、混合和扩展现实环境中的计算机视觉——文献计量学综述
IF 3.8 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2024-12-01 DOI: 10.1016/j.visinf.2024.11.002
Júlio Castro Lopes, Rui Pedro Lopes
This work describes a bibliometric analysis of the literature on the use of computer vision algorithms in Augmented Reality (AR), Virtual Reality (VR), Mixed Reality (MR), and Extended Reality (XR) environments. The analysis aims to highlight the evolution, trends, and effects of research in this field. This review provides an overview of immersive technologies and their applications, as well as the role of computer vision algorithms in enabling these technologies and the potential benefits of using such algorithms. This study identifies important authors, institutions, and research themes by using bibliometric indicators such as citation counts, co-citation analysis, and network analysis. The analysis also identifies gaps and opportunities for additional research in this area, as well as a critical assessment of the quality and relevance of the publications.
这项工作描述了在增强现实(AR)、虚拟现实(VR)、混合现实(MR)和扩展现实(XR)环境中使用计算机视觉算法的文献计量学分析。该分析旨在突出该领域研究的演变、趋势和影响。本文概述了沉浸式技术及其应用,以及计算机视觉算法在实现这些技术中的作用以及使用这些算法的潜在好处。本研究通过使用文献计量指标,如引文计数、共被引分析和网络分析,确定了重要的作者、机构和研究主题。该分析还确定了在这一领域进行进一步研究的差距和机会,并对出版物的质量和相关性进行了批判性评估。
{"title":"Computer Vision in Augmented, Virtual, Mixed and Extended Reality environments—A bibliometric review","authors":"Júlio Castro Lopes,&nbsp;Rui Pedro Lopes","doi":"10.1016/j.visinf.2024.11.002","DOIUrl":"10.1016/j.visinf.2024.11.002","url":null,"abstract":"<div><div>This work describes a bibliometric analysis of the literature on the use of computer vision algorithms in Augmented Reality (AR), Virtual Reality (VR), Mixed Reality (MR), and Extended Reality (XR) environments. The analysis aims to highlight the evolution, trends, and effects of research in this field. This review provides an overview of immersive technologies and their applications, as well as the role of computer vision algorithms in enabling these technologies and the potential benefits of using such algorithms. This study identifies important authors, institutions, and research themes by using bibliometric indicators such as citation counts, co-citation analysis, and network analysis. The analysis also identifies gaps and opportunities for additional research in this area, as well as a critical assessment of the quality and relevance of the publications.</div></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"8 4","pages":"Pages 13-22"},"PeriodicalIF":3.8,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143098854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Visual Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1