首页 > 最新文献

IEEE Computer Graphics and Applications最新文献

英文 中文
Voting-Based Intervention Planning Using AI-Generated Images. 使用人工智能生成的图像进行基于投票的干预计划。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3553620
Ioannis Kavouras, Ioannis Rallis, Emmanuel Sardis, Anastasios Doulamis, Nikolaos Doulamis

The continuous evolution of artificial intelligence and advanced algorithms capable of generating information from simplified input creates new opportunities for several scientific fields. Currently, the applicability of such technologies is limited to art and medical domains, but it can be applied to engineering domains to help the architects and urban planners design environmentally friendly solutions by proposing several alternatives in a short time. This work utilizes the image-inpainting algorithm for suggesting several alternative solutions to four European cities. In addition, this work suggests the utilization of a voting-based framework for finding the most preferred solution for each case study. The voting-based framework involves the participation of citizens and, as a result, decentralizes and democratizes the urban planning process. Finally, this research indicates the importance of deploying generative models in engineering applications by proving that generative AI models are capable of supporting the architects and urban planners in urban planning procedures.

人工智能的不断发展和先进的算法能够从简化的输入中生成信息,这为几个科学领域创造了新的机会。目前,这种技术的适用性仅限于艺术和医疗领域,但它可以应用于工程领域,通过在短时间内提出几种替代方案,帮助建筑师和城市规划者设计环境友好的解决方案。这项工作利用图像绘制算法为四个欧洲城市提供了几种替代解决方案。此外,这项工作建议利用基于投票的框架为每个案例研究找到最受欢迎的解决方案。以投票为基础的框架涉及公民的参与,从而使城市规划过程分散和民主化。最后,本研究通过证明生成人工智能模型能够在城市规划过程中支持建筑师和城市规划者,表明了在工程应用中部署生成模型的重要性。
{"title":"Voting-Based Intervention Planning Using AI-Generated Images.","authors":"Ioannis Kavouras, Ioannis Rallis, Emmanuel Sardis, Anastasios Doulamis, Nikolaos Doulamis","doi":"10.1109/MCG.2025.3553620","DOIUrl":"10.1109/MCG.2025.3553620","url":null,"abstract":"<p><p>The continuous evolution of artificial intelligence and advanced algorithms capable of generating information from simplified input creates new opportunities for several scientific fields. Currently, the applicability of such technologies is limited to art and medical domains, but it can be applied to engineering domains to help the architects and urban planners design environmentally friendly solutions by proposing several alternatives in a short time. This work utilizes the image-inpainting algorithm for suggesting several alternative solutions to four European cities. In addition, this work suggests the utilization of a voting-based framework for finding the most preferred solution for each case study. The voting-based framework involves the participation of citizens and, as a result, decentralizes and democratizes the urban planning process. Finally, this research indicates the importance of deploying generative models in engineering applications by proving that generative AI models are capable of supporting the architects and urban planners in urban planning procedures.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"31-46"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143674965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reflections on the Use of Dashboards in the COVID-19 Pandemic. 关于在COVID-19大流行期间使用仪表板的思考。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3538257
Alessio Arleo, Rita Borgo, Jorn Kohlhammer, Roy A Ruddle, H Scharlach, Xiaoru Yuan, Melanie Tory, Daniel Keefe

Dashboards have arguably been the most used visualizations during the COVID-19 pandemic. They were used to communicate its evolution to national governments for disaster mitigation, to the public domain to inform about its status, and to epidemiologists to comprehend and predict the evolution of the disease. Each design had to be tailored for different tasks and to varying audiences-in many cases set up in a very short time due to the urgent need. In this article, we collect notable examples of dashboards and reflect on their use and design during the pandemic from a user-oriented perspective. We interview a group of researchers with varying visualization expertise who actively used dashboards during the pandemic as part of their daily workflow. We discuss our findings and compile a list of lessons learned to support future visualization researchers and dashboard designers.

在COVID-19大流行期间,仪表板可以说是最常用的可视化工具。它们被用来向国家政府通报其演变情况,以减轻灾害,向公共领域通报其状况,并向流行病学家通报其演变情况,以了解和预测该疾病的演变。每个设计都必须针对不同的任务和不同的受众进行定制——在许多情况下,由于迫切需要,在很短的时间内完成。在本文中,我们收集了一些值得注意的仪表板示例,并从面向用户的角度反思它们在大流行期间的使用和设计。我们采访了一组具有不同可视化专业知识的研究人员,他们在大流行期间积极使用仪表板作为日常工作的一部分。我们讨论了我们的发现,并编制了一份经验教训清单,以支持未来的可视化研究人员和仪表板设计师。
{"title":"Reflections on the Use of Dashboards in the COVID-19 Pandemic.","authors":"Alessio Arleo, Rita Borgo, Jorn Kohlhammer, Roy A Ruddle, H Scharlach, Xiaoru Yuan, Melanie Tory, Daniel Keefe","doi":"10.1109/MCG.2025.3538257","DOIUrl":"10.1109/MCG.2025.3538257","url":null,"abstract":"<p><p>Dashboards have arguably been the most used visualizations during the COVID-19 pandemic. They were used to communicate its evolution to national governments for disaster mitigation, to the public domain to inform about its status, and to epidemiologists to comprehend and predict the evolution of the disease. Each design had to be tailored for different tasks and to varying audiences-in many cases set up in a very short time due to the urgent need. In this article, we collect notable examples of dashboards and reflect on their use and design during the pandemic from a user-oriented perspective. We interview a group of researchers with varying visualization expertise who actively used dashboards during the pandemic as part of their daily workflow. We discuss our findings and compile a list of lessons learned to support future visualization researchers and dashboard designers.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"45 2","pages":"135-142"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144287147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LossLens: Diagnostics for Machine Learning Through Loss Landscape Visual Analytics. LossLens:通过损失景观视觉分析进行机器学习诊断。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2024.3509374
Tiankai Xie, Jiaqing Chen, Yaoqing Yang, Caleb Geniesse, Ge Shi, Ajinkya Jeevan Chaudhari, John Kevin Cava, Michael W Mahoney, Talita Perciano, Gunther H Weber, Ross Maciejewski

Modern machine learning often relies on optimizing a neural network's parameters using a loss function to learn complex features. Beyond training, examining the loss function with respect to a network's parameters (i.e., as a loss landscape) can reveal insights into the architecture and learning process. While the local structure of the loss landscape surrounding an individual solution can be characterized using a variety of approaches, the global structure of a loss landscape, which includes potentially many local minima corresponding to different solutions, remains far more difficult to conceptualize and visualize. To address this difficulty, we introduce LossLens, a visual analytics framework that explores loss landscapes at multiple scales. LossLens integrates metrics from global and local scales into a comprehensive visual representation, enhancing model diagnostics. We demonstrate LossLens through two case studies: visualizing how residual connections influence a ResNet-20, and visualizing how physical parameters influence a physics-informed neural network solving a simple convection problem.

现代机器学习通常依赖于使用损失函数来优化神经网络的参数来学习复杂的特征。除了训练之外,检查网络参数的损失函数(即,作为损失景观)可以揭示对架构和学习过程的见解。虽然围绕单个解决方案的损失格局的局部结构可以使用各种方法来表征,但损失格局的整体结构,其中可能包括与不同解决方案对应的许多局部最小值,仍然难以概念化和可视化。为了解决这个问题,我们引入了LossLens,这是一个视觉分析框架,可以在多个尺度上探索损失景观。LossLens将来自全球和局部尺度的指标集成到全面的视觉表示中,增强了模型诊断。我们通过两个案例研究来演示LossLens:可视化剩余连接如何影响ResNet-20,以及可视化物理参数如何影响物理信息神经网络(PINN)解决简单对流问题。
{"title":"LossLens: Diagnostics for Machine Learning Through Loss Landscape Visual Analytics.","authors":"Tiankai Xie, Jiaqing Chen, Yaoqing Yang, Caleb Geniesse, Ge Shi, Ajinkya Jeevan Chaudhari, John Kevin Cava, Michael W Mahoney, Talita Perciano, Gunther H Weber, Ross Maciejewski","doi":"10.1109/MCG.2024.3509374","DOIUrl":"10.1109/MCG.2024.3509374","url":null,"abstract":"<p><p>Modern machine learning often relies on optimizing a neural network's parameters using a loss function to learn complex features. Beyond training, examining the loss function with respect to a network's parameters (i.e., as a loss landscape) can reveal insights into the architecture and learning process. While the local structure of the loss landscape surrounding an individual solution can be characterized using a variety of approaches, the global structure of a loss landscape, which includes potentially many local minima corresponding to different solutions, remains far more difficult to conceptualize and visualize. To address this difficulty, we introduce LossLens, a visual analytics framework that explores loss landscapes at multiple scales. LossLens integrates metrics from global and local scales into a comprehensive visual representation, enhancing model diagnostics. We demonstrate LossLens through two case studies: visualizing how residual connections influence a ResNet-20, and visualizing how physical parameters influence a physics-informed neural network solving a simple convection problem.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"112-125"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143544540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HoloJig: Interactive Spoken Prompt Specified Generative AI Environments. HoloJig:交互式语音提示指定生成AI环境。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3553780
Llogari Casas, Samantha Hannah, Kenny Mitchell

HoloJig offers an interactive, speech-to-virtual reality (VR), VR experience that generates diverse environments in real time based on live spoken descriptions. Unlike traditional VR systems that rely on prebuilt assets, HoloJig dynamically creates personalized and immersive virtual spaces with depth-based parallax 3-D rendering, allowing users to define the characteristics of their immersive environment through verbal prompts. This generative approach opens up new possibilities for interactive experiences, including simulations, training, collaborative workspaces, and entertainment. In addition to speech-to-VR environment generation, a key innovation of HoloJig is its progressive visual transition mechanism, which smoothly dissolves between previously generated and newly requested environments, mitigating the delay caused by neural computations. This feature ensures a seamless and continuous user experience, even as new scenes are being rendered on remote servers.

HoloJig提供了一种交互式的语音到虚拟现实的虚拟现实体验,可以根据现场语音描述实时生成各种环境。与依赖于预构建资产的传统VR系统不同,HoloJig通过基于深度的视差3D渲染动态创建个性化和沉浸式虚拟空间,允许用户通过口头提示定义沉浸式环境的特征。这种生成方法为互动体验开辟了新的可能性,包括模拟、培训、协作工作空间和娱乐。除了语音到vr环境的生成,HoloJig的一个关键创新是其渐进式视觉转换机制,该机制可以在先前生成的环境和新请求的环境之间平滑地分解,减轻了神经计算造成的延迟。该功能确保了无缝和连续的用户体验,即使是在远程服务器上呈现新场景。
{"title":"HoloJig: Interactive Spoken Prompt Specified Generative AI Environments.","authors":"Llogari Casas, Samantha Hannah, Kenny Mitchell","doi":"10.1109/MCG.2025.3553780","DOIUrl":"10.1109/MCG.2025.3553780","url":null,"abstract":"<p><p>HoloJig offers an interactive, speech-to-virtual reality (VR), VR experience that generates diverse environments in real time based on live spoken descriptions. Unlike traditional VR systems that rely on prebuilt assets, HoloJig dynamically creates personalized and immersive virtual spaces with depth-based parallax 3-D rendering, allowing users to define the characteristics of their immersive environment through verbal prompts. This generative approach opens up new possibilities for interactive experiences, including simulations, training, collaborative workspaces, and entertainment. In addition to speech-to-VR environment generation, a key innovation of HoloJig is its progressive visual transition mechanism, which smoothly dissolves between previously generated and newly requested environments, mitigating the delay caused by neural computations. This feature ensures a seamless and continuous user experience, even as new scenes are being rendered on remote servers.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"69-77"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143674960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Should I Render or Should AI Generate? Crafting Synthetic Semantic Segmentation Datasets With Controlled Generation. 我应该渲染还是AI生成?合成语义分割数据集与控制生成。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3553494
Omar A Mures, Manuel Silva, Manuel Lijo-Sanchez, Emilio J Padron, Jose A Iglesias-Guitian

This work explores the integration of generative AI models for automatically generating synthetic image-labeled data. Our approach leverages controllable diffusion models to generate synthetic variations of semantically labeled images. Synthetic datasets for semantic segmentation struggle to represent real-world subtleties, such as different weather conditions or fine details, typically relying on costly simulations and rendering. However, diffusion models can generate diverse images using input text prompts and guidance images, such as semantic masks. Our work introduces and tests a novel methodology for generating labeled synthetic images, with an initial focus on semantic segmentation, a demanding computer vision task. We showcase our approach in two distinct image segmentation domains, outperforming traditional computer graphics simulations in efficiently creating diverse datasets and training downstream models. We leverage generative models for crafting synthetically labeled images, posing the question: "Should I render or should AI generate?" Our results endorse a paradigm shift toward controlled generation models.

这项工作探讨了如何整合生成式人工智能模型,以自动生成合成图像标签数据。我们的方法利用可控扩散模型生成语义标签图像的合成变化。用于语义分割的合成数据集很难表现真实世界的微妙之处,如不同的天气条件或精细细节,通常依赖于昂贵的模拟和渲染。然而,扩散模型可以利用输入文本提示和引导图像(如语义掩码)生成不同的图像。我们的工作介绍并测试了一种生成带标签合成图像的新方法,最初的重点是语义分割,这是一项要求很高的计算机视觉任务。我们在两个不同的图像分割领域展示了我们的方法,在高效创建各种数据集和训练下游模型方面,我们的方法优于传统的计算机图形模拟。我们利用生成模型制作合成标签图像,提出了一个问题:"我应该渲染还是人工智能应该生成?我们的研究结果支持向受控生成模型的范式转变。
{"title":"Should I Render or Should AI Generate? Crafting Synthetic Semantic Segmentation Datasets With Controlled Generation.","authors":"Omar A Mures, Manuel Silva, Manuel Lijo-Sanchez, Emilio J Padron, Jose A Iglesias-Guitian","doi":"10.1109/MCG.2025.3553494","DOIUrl":"10.1109/MCG.2025.3553494","url":null,"abstract":"<p><p>This work explores the integration of generative AI models for automatically generating synthetic image-labeled data. Our approach leverages controllable diffusion models to generate synthetic variations of semantically labeled images. Synthetic datasets for semantic segmentation struggle to represent real-world subtleties, such as different weather conditions or fine details, typically relying on costly simulations and rendering. However, diffusion models can generate diverse images using input text prompts and guidance images, such as semantic masks. Our work introduces and tests a novel methodology for generating labeled synthetic images, with an initial focus on semantic segmentation, a demanding computer vision task. We showcase our approach in two distinct image segmentation domains, outperforming traditional computer graphics simulations in efficiently creating diverse datasets and training downstream models. We leverage generative models for crafting synthetically labeled images, posing the question: \"Should I render or should AI generate?\" Our results endorse a paradigm shift toward controlled generation models.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"57-68"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143674963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unified Visual Comparison Framework for Human and AI Paintings Using Neural Embeddings and Computational Aesthetics. 使用神经嵌入和计算美学的人类和人工智能绘画的统一视觉比较框架。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3555122
Yilin Ye, Rong Huang, Kang Zhang, Wei Zeng

To facilitate comparative analysis of artificial intelligence (AI) and human paintings, we present a unified computational framework combining neural embedding and computational aesthetic features. We first exploit CLIP embedding to provide a projected overview for human and AI painting datasets, and we next leverage computational aesthetic metrics to obtain explainable features of paintings. On that basis, we design a visual analytics system that involves distribution discrepancy measurement for quantifying dataset differences and evolutionary analysis for comparing artists with AI models. Case studies comparing three AI-generated datasets with three human paintings datasets, and analyzing the evolutionary differences between authentic Picasso paintings and AI-generated ones, show the effectiveness of our framework.

为了便于人工智能和人类绘画的比较分析,我们提出了一个结合神经嵌入和计算美学特征的统一计算框架。我们首先利用CLIP嵌入来提供人类和人工智能绘画数据集的投影概述,然后利用计算美学指标来获得绘画的可解释特征。在此基础上,我们设计了一个可视化分析系统,该系统包括用于量化数据集差异的分布差异测量和用于比较艺术家与AI模型的进化分析。案例研究将三个人工智能生成的数据集与三个人类绘画数据集进行比较,并分析毕加索真迹与人工智能生成的画作之间的进化差异,表明了我们的框架的有效性。
{"title":"Unified Visual Comparison Framework for Human and AI Paintings Using Neural Embeddings and Computational Aesthetics.","authors":"Yilin Ye, Rong Huang, Kang Zhang, Wei Zeng","doi":"10.1109/MCG.2025.3555122","DOIUrl":"10.1109/MCG.2025.3555122","url":null,"abstract":"<p><p>To facilitate comparative analysis of artificial intelligence (AI) and human paintings, we present a unified computational framework combining neural embedding and computational aesthetic features. We first exploit CLIP embedding to provide a projected overview for human and AI painting datasets, and we next leverage computational aesthetic metrics to obtain explainable features of paintings. On that basis, we design a visual analytics system that involves distribution discrepancy measurement for quantifying dataset differences and evolutionary analysis for comparing artists with AI models. Case studies comparing three AI-generated datasets with three human paintings datasets, and analyzing the evolutionary differences between authentic Picasso paintings and AI-generated ones, show the effectiveness of our framework.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"19-30"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143765960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visual Analytics for Explainable and Trustworthy Artificial Intelligence. 可解释和可信赖的人工智能的可视化分析。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3533806
Angelos Chatzimparmpas, Sumanta N Pattanaik

Our society increasingly depends on intelligent systems to solve complex problems, ranging from recommender systems suggesting the next movie to watch to AI models assisting in medical diagnoses for hospitalized patients. With the iterative improvement of diagnostic accuracy and efficiency, AI holds significant potential to mitigate medical misdiagnoses by preventing numerous deaths and reducing an economic burden of approximately € 450 billion annually. However, a key obstacle to AI adoption lies in the lack of transparency, that is, many automated systems provide predictions without revealing the underlying processes. This opacity can hinder experts' ability to trust and rely on AI systems. Visual analytics (VA) provides a compelling solution by combining AI models with interactive visualizations. These specialized charts and graphs empower users to incorporate their domain expertise to refine and improve the models, bridging the gap between AI and human understanding. In this work, the author defines, categorizes, and explores how VA solutions can foster trust across the stages of a typical AI pipeline. The author proposes a design space for innovative visualizations and presents an overview of our previously developed VA dashboards, which support critical tasks within the various pipeline stages, including data processing, feature engineering, hyperparameter tuning, understanding, debugging, refining, and comparing models.

我们的社会越来越依赖智能系统来解决复杂的问题,从推荐下一部电影的推荐系统到帮助住院病人进行医疗诊断的人工智能模型。随着诊断准确性和效率的不断提高,人工智能具有巨大的潜力,可以通过防止大量死亡和减少每年约4500亿欧元的经济负担来减轻医疗误诊。然而,人工智能应用的一个关键障碍在于缺乏透明度,也就是说,许多自动化系统提供的预测没有揭示潜在的过程。这种不透明性可能会阻碍专家信任和依赖人工智能系统的能力。可视化分析(VA)通过将人工智能模型与交互式可视化相结合,提供了一个引人注目的解决方案。这些专门的图表和图形使用户能够结合他们的领域专业知识来完善和改进模型,弥合人工智能和人类理解之间的差距。在这项工作中,作者定义、分类并探讨了VA解决方案如何在典型的人工智能管道的各个阶段促进信任。作者提出了一个创新可视化的设计空间,并概述了我们以前开发的VA仪表板,它支持各个管道阶段中的关键任务,包括数据处理、特征工程、超参数调优、理解、调试、精炼和比较模型。
{"title":"Visual Analytics for Explainable and Trustworthy Artificial Intelligence.","authors":"Angelos Chatzimparmpas, Sumanta N Pattanaik","doi":"10.1109/MCG.2025.3533806","DOIUrl":"10.1109/MCG.2025.3533806","url":null,"abstract":"<p><p>Our society increasingly depends on intelligent systems to solve complex problems, ranging from recommender systems suggesting the next movie to watch to AI models assisting in medical diagnoses for hospitalized patients. With the iterative improvement of diagnostic accuracy and efficiency, AI holds significant potential to mitigate medical misdiagnoses by preventing numerous deaths and reducing an economic burden of approximately € 450 billion annually. However, a key obstacle to AI adoption lies in the lack of transparency, that is, many automated systems provide predictions without revealing the underlying processes. This opacity can hinder experts' ability to trust and rely on AI systems. Visual analytics (VA) provides a compelling solution by combining AI models with interactive visualizations. These specialized charts and graphs empower users to incorporate their domain expertise to refine and improve the models, bridging the gap between AI and human understanding. In this work, the author defines, categorizes, and explores how VA solutions can foster trust across the stages of a typical AI pipeline. The author proposes a design space for innovative visualizations and presents an overview of our previously developed VA dashboards, which support critical tasks within the various pipeline stages, including data processing, feature engineering, hyperparameter tuning, understanding, debugging, refining, and comparing models.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"45 2","pages":"100-111"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144287149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mixed Reality Collaboration: How Expert Representations Shape User Experiences. 混合现实协作:专家代表如何塑造用户体验。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3537316
Bernardo Marques, Samuel Silva, Carlos Ferreira, Sergio Oliveira, Andreia Santos, Paulo Dias, Beatriz Sousa Santos, Mike Potel

Remote assistance has become integral in today's work environments, enabling offsite experts to guide team members in need of assistance. One technology that has gained momentum is mixed reality (MR), combining virtual and augmented reality elements, bridging physical distances, and facilitating knowledge sharing. One critical area is the representation of remote experts. Understanding how different representations influence the learning process is essential for improving the shared experience. Effective representation can enhance communication, trust, and the perceived expertise of the remote expert, thereby improving the learning and collaboration experiences. This study focuses on evaluating how different representations affect various dimensions of collaboration. The study involved 57 participants and examined six different conditions: no visual representation, representation using emojis, representation using a cartoon avatar, representation using a robot avatar, representation using a realistic avatar, and live stream using video.

远程协助已成为当今工作环境中不可或缺的一部分,使远程专家能够指导需要帮助的团队成员。一项发展势头强劲的技术是混合现实(MR),它结合了虚拟现实和增强现实元素,弥合了物理距离,促进了知识共享。一个关键领域是远程专家的代表。了解不同表征如何影响学习过程对于改善共享体验至关重要。有效的代表可以增强远程专家的沟通、信任和感知的专业知识,从而改善学习和协作体验。本研究的重点是评估不同的表征如何影响合作的各个维度。这项研究涉及57名参与者,并检查了六种不同的情况:没有视觉表现,使用表情符号表现,使用卡通化身表现,使用机器人化身表现,使用现实化身表现,以及使用视频直播。
{"title":"Mixed Reality Collaboration: How Expert Representations Shape User Experiences.","authors":"Bernardo Marques, Samuel Silva, Carlos Ferreira, Sergio Oliveira, Andreia Santos, Paulo Dias, Beatriz Sousa Santos, Mike Potel","doi":"10.1109/MCG.2025.3537316","DOIUrl":"https://doi.org/10.1109/MCG.2025.3537316","url":null,"abstract":"<p><p>Remote assistance has become integral in today's work environments, enabling offsite experts to guide team members in need of assistance. One technology that has gained momentum is mixed reality (MR), combining virtual and augmented reality elements, bridging physical distances, and facilitating knowledge sharing. One critical area is the representation of remote experts. Understanding how different representations influence the learning process is essential for improving the shared experience. Effective representation can enhance communication, trust, and the perceived expertise of the remote expert, thereby improving the learning and collaboration experiences. This study focuses on evaluating how different representations affect various dimensions of collaboration. The study involved 57 participants and examined six different conditions: no visual representation, representation using emojis, representation using a cartoon avatar, representation using a robot avatar, representation using a realistic avatar, and live stream using video.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"45 2","pages":"143-151"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144287146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Meet-in-Style: Text-Driven Real-Time Video Stylization Using Diffusion Models. 风格相遇:使用扩散模型的文本驱动的实时视频风格化。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3554312
David Kunz, Ondrej Texler, David Mould, Daniel Sykora

We present Meet-in-Style-a new approach to real-time stylization of live video streams using text prompts. In contrast to previous text-based techniques, our system is able to stylize input video at 30 fps on commodity graphics hardware while preserving structural consistency of the stylized sequence and minimizing temporal flicker. A key idea of our approach is to combine diffusion-based image stylization with a few-shot patch-based training strategy that can produce a custom image-to-image stylization network with real-time inference capabilities. Such a combination not only allows for fast stylization, but also greatly improves consistency of individual stylized frames compared to a scenario where diffusion is applied to each video frame separately. We conducted a number of user experiments in which we found our approach to be particularly useful in video conference scenarios enabling participants to interactively apply different visual styles to themselves (or to each other) to enhance the overall chatting experience.

我们提出meet - in - style -一种使用文本提示的实时视频流样式化的新方法。与之前基于文本的技术相比,我们的系统能够在商用图形硬件上以30 fps的速度对输入视频进行风格化,同时保持风格化序列的结构一致性并最大限度地减少时间闪烁。我们方法的一个关键思想是将基于扩散的图像风格化与基于少量镜头补丁的训练策略相结合,该策略可以产生具有实时推理能力的自定义图像到图像风格化网络。这样的组合不仅允许快速风格化,而且与单独应用扩散到每个视频帧的场景相比,还大大提高了单个风格化帧的一致性。我们进行了大量的用户实验,发现我们的方法在视频会议场景中特别有用,使参与者能够交互式地为自己(或彼此)应用不同的视觉风格,以增强整体聊天体验。
{"title":"Meet-in-Style: Text-Driven Real-Time Video Stylization Using Diffusion Models.","authors":"David Kunz, Ondrej Texler, David Mould, Daniel Sykora","doi":"10.1109/MCG.2025.3554312","DOIUrl":"10.1109/MCG.2025.3554312","url":null,"abstract":"<p><p>We present Meet-in-Style-a new approach to real-time stylization of live video streams using text prompts. In contrast to previous text-based techniques, our system is able to stylize input video at 30 fps on commodity graphics hardware while preserving structural consistency of the stylized sequence and minimizing temporal flicker. A key idea of our approach is to combine diffusion-based image stylization with a few-shot patch-based training strategy that can produce a custom image-to-image stylization network with real-time inference capabilities. Such a combination not only allows for fast stylization, but also greatly improves consistency of individual stylized frames compared to a scenario where diffusion is applied to each video frame separately. We conducted a number of user experiments in which we found our approach to be particularly useful in video conference scenarios enabling participants to interactively apply different visual styles to themselves (or to each other) to enhance the overall chatting experience.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"47-56"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143702373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Trame: Platform Ubiquitous, Scalable Integration Framework for Visual Analytics. 框架:平台无处不在,可扩展的可视化分析集成框架。
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-03-01 DOI: 10.1109/MCG.2025.3540264
Se'bastien Jourdain, Patrick O'Leary, Will Schroeder, Nicholas F Polys

Trame is an open-source, Python-based, scalable integration framework for visual analytics. It is the culmination of decades of work-by a large and active community-beginning with the creation of VTK, the growth of ParaView as a premier high-performance, client-server computing system, and more recently the creation of web tools, such as VTK.js and VTK.wasm. As an integration environment, trame relies on open-source standards and tools that can be easily combined into effective computing solutions. We have long recognized that impactful analytics tools must be ubiquitous-meaning they run on all major computing platforms-and integrate/interoperate easily with external packages, such as data systems and processing tools, application UI frameworks, and 2-D/3-D graphical libraries. In this article, we present the architecture and use of trame for applications ranging from simple dashboards to complex workflow-based applications. We also describe examples that readily incorporate external tools and run without coding changes on desktop, mobile, cloud, client-server, and interactive computing notebooks, such as Jupyter.

frame是一个开源的、基于python的、可扩展的可视化分析集成框架。它是一个庞大而活跃的社区几十年工作的高潮——从创建VTK开始,ParaView作为一个高性能的客户端-服务器计算系统的发展,以及最近创建的web工具,如VTK.js和VTK.wasm。作为一个集成环境,frame依赖于开源标准和工具,这些标准和工具可以很容易地组合成有效的计算解决方案。我们早就认识到,有影响力的分析工具必须无处不在——这意味着它们可以在所有主要的计算平台上运行——并且可以轻松地与外部包(如数据系统和处理工具、应用程序UI框架和2-D/3-D图形库)集成/互操作。在本文中,我们将介绍从简单的仪表板到复杂的基于工作流的应用程序的框架的体系结构和使用。我们还描述了一些示例,这些示例可以轻松地合并外部工具并在桌面、移动设备、云、客户机-服务器和交互式计算笔记本(如Jupyter)上无需更改代码即可运行。
{"title":"Trame: Platform Ubiquitous, Scalable Integration Framework for Visual Analytics.","authors":"Se'bastien Jourdain, Patrick O'Leary, Will Schroeder, Nicholas F Polys","doi":"10.1109/MCG.2025.3540264","DOIUrl":"https://doi.org/10.1109/MCG.2025.3540264","url":null,"abstract":"<p><p>Trame is an open-source, Python-based, scalable integration framework for visual analytics. It is the culmination of decades of work-by a large and active community-beginning with the creation of VTK, the growth of ParaView as a premier high-performance, client-server computing system, and more recently the creation of web tools, such as VTK.js and VTK.wasm. As an integration environment, trame relies on open-source standards and tools that can be easily combined into effective computing solutions. We have long recognized that impactful analytics tools must be ubiquitous-meaning they run on all major computing platforms-and integrate/interoperate easily with external packages, such as data systems and processing tools, application UI frameworks, and 2-D/3-D graphical libraries. In this article, we present the architecture and use of trame for applications ranging from simple dashboards to complex workflow-based applications. We also describe examples that readily incorporate external tools and run without coding changes on desktop, mobile, cloud, client-server, and interactive computing notebooks, such as Jupyter.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"45 2","pages":"126-134"},"PeriodicalIF":1.7,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144287148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Computer Graphics and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1