首页 > 最新文献

IEEE Computer Graphics and Applications最新文献

英文 中文
How Visually Literate Are Large Language Models? Reflections on Recent Advances and Future Directions. 大型语言模型的视觉素养如何?对最近进展和未来方向的思考。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-01 DOI: 10.1109/MCG.2025.3605029
Alexander Bendeck, John Stasko, Rahul C Basole, Francesco Ferrise

Large language models (LLMs) are now being applied to the tasks of visualization generation and understanding, demonstrating these models' ability to be "visually literate." On the generation side, LLMs have shown promise in powering natural languages' interfaces for visualization authoring while also suffering from usability and inconsistency issues. On the interpretation side, models (especially vision-language models) can answer basic questions about visualizations, synthesize visual and textual information, and detect misleading visual designs. However, models also tend to struggle with certain analytic tasks, and their takeaways from reading visualizations often differ from those of humans. We aim to both illuminate the state of the art in LLMs' visualization literacy and speculate on where such work may, and perhaps ought to, take us next.

大型语言模型(llm)现在被应用于可视化生成和理解的任务,展示了这些模型具有“视觉素养”的能力。在生成方面,法学硕士在为可视化创作提供自然语言接口方面表现出了希望,同时也受到可用性和不一致性问题的困扰。在解释方面,模型(尤其是视觉语言模型)可以回答关于可视化的基本问题,综合视觉和文本信息,并检测误导性的视觉设计。然而,模型也倾向于与某些分析任务作斗争,并且它们从阅读可视化中得到的结论通常与人类不同。我们的目标是阐明法学硕士可视化素养的现状,并推测此类工作可能(或许应该)将我们带往何处。
{"title":"How Visually Literate Are Large Language Models? Reflections on Recent Advances and Future Directions.","authors":"Alexander Bendeck, John Stasko, Rahul C Basole, Francesco Ferrise","doi":"10.1109/MCG.2025.3605029","DOIUrl":"10.1109/MCG.2025.3605029","url":null,"abstract":"<p><p>Large language models (LLMs) are now being applied to the tasks of visualization generation and understanding, demonstrating these models' ability to be \"visually literate.\" On the generation side, LLMs have shown promise in powering natural languages' interfaces for visualization authoring while also suffering from usability and inconsistency issues. On the interpretation side, models (especially vision-language models) can answer basic questions about visualizations, synthesize visual and textual information, and detect misleading visual designs. However, models also tend to struggle with certain analytic tasks, and their takeaways from reading visualizations often differ from those of humans. We aim to both illuminate the state of the art in LLMs' visualization literacy and speculate on where such work may, and perhaps ought to, take us next.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"45 6","pages":"120-129"},"PeriodicalIF":1.4,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145497505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MuCHEx: A Multimodal Conversational Debugging Tool for Interactive Visual Exploration of Hierarchical Object Classification. 多级对象分类的交互式可视化探索的多模态会话调试工具。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-01 DOI: 10.1109/MCG.2025.3598204
Reza Shahriari, Yichi Yang, Danish Nisar Ahmed Tamboli, Michael Perez, Yuheng Zha, Jinyu Hou, Mingkai Deng, Eric D Ragan, Jaime Ruiz, Daisy Zhe Wang, Zhiting Hu, Eric Xing

Object recognition is a fundamental challenge in computer vision, particularly for fine-grained object classification, where classes differ in minor features. Improved fine-grained object classification requires a teaching system with numerous classes and instances of data. As the number of hierarchical levels and instances grows, debugging these models becomes increasingly complex. Moreover, different types of debugging tasks require varying approaches, explanations, and levels of detail. We present MuCHEx, a multimodal conversational system that blends natural language and visual interaction for interactive debugging of hierarchical object classification. Natural language allows users to flexibly express high-level questions or debugging goals without needing to navigate complex interfaces, while adaptive explanations surface only the most relevant visual or textual details based on the user's current task. This multimodal approach combines the expressiveness of language with the precision of direct manipulation, enabling context-aware exploration during model debugging.

对象识别是计算机视觉中的一个基本挑战,特别是对于细粒度对象分类,其中类在次要特征上有所不同。改进的细粒度对象分类需要一个包含大量类和数据实例的教学系统。随着分层级别和实例数量的增长,调试这些模型变得越来越复杂。此外,不同类型的调试任务需要不同的方法、解释和详细程度。我们提出了MuCHEx,一个混合了自然语言和视觉交互的多模态会话系统,用于分层对象分类的交互式调试。自然语言允许用户灵活地表达高级问题或调试目标,而无需导航复杂的界面,而自适应解释仅根据用户当前任务显示最相关的视觉或文本细节。这种多模态方法结合了语言的表现力和直接操作的精确性,在模型调试期间实现了上下文感知的探索。
{"title":"MuCHEx: A Multimodal Conversational Debugging Tool for Interactive Visual Exploration of Hierarchical Object Classification.","authors":"Reza Shahriari, Yichi Yang, Danish Nisar Ahmed Tamboli, Michael Perez, Yuheng Zha, Jinyu Hou, Mingkai Deng, Eric D Ragan, Jaime Ruiz, Daisy Zhe Wang, Zhiting Hu, Eric Xing","doi":"10.1109/MCG.2025.3598204","DOIUrl":"10.1109/MCG.2025.3598204","url":null,"abstract":"<p><p>Object recognition is a fundamental challenge in computer vision, particularly for fine-grained object classification, where classes differ in minor features. Improved fine-grained object classification requires a teaching system with numerous classes and instances of data. As the number of hierarchical levels and instances grows, debugging these models becomes increasingly complex. Moreover, different types of debugging tasks require varying approaches, explanations, and levels of detail. We present MuCHEx, a multimodal conversational system that blends natural language and visual interaction for interactive debugging of hierarchical object classification. Natural language allows users to flexibly express high-level questions or debugging goals without needing to navigate complex interfaces, while adaptive explanations surface only the most relevant visual or textual details based on the user's current task. This multimodal approach combines the expressiveness of language with the precision of direct manipulation, enabling context-aware exploration during model debugging.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"76-88"},"PeriodicalIF":1.4,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144838637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AnchorTextVis: A Visual Analytics Approach for Fast Comparison of Text Embeddings. 一个快速比较文本嵌入的可视化分析方法。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-01 DOI: 10.1109/MCG.2025.3598262
Jingzhen Zhang, Hongjiang Lv, Zhibin Niu

Visual comparison of text embeddings is crucial for analyzing semantic differences and comparing embedding models. Existing methods fail to maintain visual consistency in comparative regions and lack AI-assisted analysis, leading to high cognitive loads and time-consuming exploration processes. In this article, we propose AnchorTextVis, a visual analytics approach based on AnchorMap-our dynamic projection algorithm balancing spatial quality and temporal coherence and large language models (LLMs) to preserve users' mental map and accelerate the exploration process. We introduce the use of comparable dimensionality reduction algorithms that maintain visual consistency, such as AnchorMap from our previous work and Joint t-SNE. Building on this foundation, we leverage LLMs to compare and summarize, offering users insights. For quantitative comparisons, we define two complementary metrics, Shared k-nearest neighbors (KNN) and Coordinate distance. Besides, we have also designed intuitive representation and rich interactive tools to compare clusters of texts and individual texts. We demonstrate the effectiveness and usefulness of our approach through three case studies and expert feedback.

文本嵌入的视觉比较是分析语义差异和比较嵌入模型的关键。现有方法无法保持比较区域的视觉一致性,并且缺乏人工智能辅助分析,导致高认知负荷和耗时的探索过程。在本文中,我们提出了一种基于anchormap的可视化分析方法——我们的动态投影算法,平衡空间质量和时间一致性,以及llm,以保留用户的心理地图并加速探索过程。我们介绍了保持视觉一致性的可比较的降维算法的使用,例如我们之前工作中的AnchorMap和Joint t-SNE。在此基础上,我们利用法学硕士进行比较和总结,为用户提供见解。为了进行定量比较,我们定义了两个互补的度量,共享KNN和坐标距离。此外,我们还设计了直观的表示和丰富的交互工具来比较文本簇和单个文本。我们通过三个案例研究和专家反馈证明了我们方法的有效性和实用性。
{"title":"AnchorTextVis: A Visual Analytics Approach for Fast Comparison of Text Embeddings.","authors":"Jingzhen Zhang, Hongjiang Lv, Zhibin Niu","doi":"10.1109/MCG.2025.3598262","DOIUrl":"10.1109/MCG.2025.3598262","url":null,"abstract":"<p><p>Visual comparison of text embeddings is crucial for analyzing semantic differences and comparing embedding models. Existing methods fail to maintain visual consistency in comparative regions and lack AI-assisted analysis, leading to high cognitive loads and time-consuming exploration processes. In this article, we propose AnchorTextVis, a visual analytics approach based on AnchorMap-our dynamic projection algorithm balancing spatial quality and temporal coherence and large language models (LLMs) to preserve users' mental map and accelerate the exploration process. We introduce the use of comparable dimensionality reduction algorithms that maintain visual consistency, such as AnchorMap from our previous work and Joint t-SNE. Building on this foundation, we leverage LLMs to compare and summarize, offering users insights. For quantitative comparisons, we define two complementary metrics, Shared k-nearest neighbors (KNN) and Coordinate distance. Besides, we have also designed intuitive representation and rich interactive tools to compare clusters of texts and individual texts. We demonstrate the effectiveness and usefulness of our approach through three case studies and expert feedback.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"29-43"},"PeriodicalIF":1.4,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144849685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Agentic Visualization: Extracting Agent-Based Design Patterns From Visualization Systems. 代理可视化:从可视化系统中提取基于代理的设计模式。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-01 DOI: 10.1109/MCG.2025.3607741
Vaishali Dhanoa, Anton Wolter, Gabriela Molina Leon, Hans-Jorg Schulz, Niklas Elmqvist

Autonomous agents powered by large language models are transforming artificial intelligence (AI), creating an imperative for the visualization area. However, our field's focus on a human in the sensemaking loop raises critical questions about autonomy, delegation, and coordination for such agentic visualization that preserve human agency while amplifying analytical capabilities. This article addresses these questions by reinterpreting existing visualization systems with semiautomated or fully automatic AI components through an agentic lens. Based on this analysis, we extract a collection of design patterns for agentic visualization, including agentic roles, communication, and coordination. These patterns provide a foundation for future agentic visualization systems that effectively harness AI agents while maintaining human insight and control.

由大型语言模型驱动的自主代理正在改变人工智能,为可视化领域创造了一个必要条件。然而,我们的领域关注的是人在语义制造循环中提出了关于自主性、授权和协调的关键问题,这种代理可视化在增强分析能力的同时保留了人的代理。本文通过通过代理透镜重新解释现有的具有半自动或全自动AI组件的可视化系统来解决这些问题。在此基础上,我们提取了一组代理可视化的设计模式,包括代理角色、沟通和协调。这些模式为未来的代理可视化系统提供了基础,这些系统可以有效地利用人工智能代理,同时保持人类的洞察力和控制力。
{"title":"Agentic Visualization: Extracting Agent-Based Design Patterns From Visualization Systems.","authors":"Vaishali Dhanoa, Anton Wolter, Gabriela Molina Leon, Hans-Jorg Schulz, Niklas Elmqvist","doi":"10.1109/MCG.2025.3607741","DOIUrl":"10.1109/MCG.2025.3607741","url":null,"abstract":"<p><p>Autonomous agents powered by large language models are transforming artificial intelligence (AI), creating an imperative for the visualization area. However, our field's focus on a human in the sensemaking loop raises critical questions about autonomy, delegation, and coordination for such agentic visualization that preserve human agency while amplifying analytical capabilities. This article addresses these questions by reinterpreting existing visualization systems with semiautomated or fully automatic AI components through an agentic lens. Based on this analysis, we extract a collection of design patterns for agentic visualization, including agentic roles, communication, and coordination. These patterns provide a foundation for future agentic visualization systems that effectively harness AI agents while maintaining human insight and control.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"89-100"},"PeriodicalIF":1.4,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145031020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MidSurfer: A Parameter-Free Approach for Mid-Surface Extraction From Segmented Volumetric Data. MidSurfer:一种从分割体数据中提取中表面的无参数方法。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-23 DOI: 10.1109/MCG.2025.3624572
Eva Bones, Dawar Khan, Ciril Bohak, Benjamin A Barad, Danielle A Grotjahn, Ivan Viola, Thomas Theusl

This paper presents MidSurfer, a novel parameter-free method for extracting mid-surfaces from segmented volumetric data. The method generates uniformly triangulated, smooth meshes that accurately capture structural features. The process begins with the Ridge Field Transformation step that transforms the segmented input data, followed by the Mid-Polyline Extraction Algorithm that works on individual volume slices. Based on the connectivity of components, this step can result in either single or multiple polyline segments that represent the structural features. These segments form a coherent series, creating a backbone of regularly spaced points representing the mid-surface. Subsequently, we employ a Polyline Zipper Algorithm for triangulation that connects these polyline segments across neighboring slices, yielding a detailed triangulated mid-surface mesh. Results show that this method outperforms previous techniques in versatility, simplicity, and accuracy. Our approach is publicly available as a ParaView plugin at https://github.com/kaust-vislab/MidSurfer.

提出了一种从分割体数据中提取中间曲面的无参数方法MidSurfer。该方法生成均匀的三角化平滑网格,准确捕获结构特征。该过程从转换分割输入数据的Ridge Field Transformation步骤开始,然后是在单个体积切片上工作的Mid-Polyline Extraction Algorithm。基于组件的连通性,这一步可以产生单个或多个表示结构特征的多线段。这些部分形成了一个连贯的系列,形成了一个代表中表面的规则间隔点的主干。随后,我们采用一种多线段拉链算法(Polyline zippers Algorithm)进行三角剖分,将这些多线段跨相邻切片连接起来,生成详细的三角剖分中表面网格。结果表明,该方法在通用性、简单性和准确性方面优于以往的方法。我们的方法可以在https://github.com/kaust-vislab/MidSurfer上作为ParaView插件公开获得。
{"title":"MidSurfer: A Parameter-Free Approach for Mid-Surface Extraction From Segmented Volumetric Data.","authors":"Eva Bones, Dawar Khan, Ciril Bohak, Benjamin A Barad, Danielle A Grotjahn, Ivan Viola, Thomas Theusl","doi":"10.1109/MCG.2025.3624572","DOIUrl":"https://doi.org/10.1109/MCG.2025.3624572","url":null,"abstract":"<p><p>This paper presents MidSurfer, a novel parameter-free method for extracting mid-surfaces from segmented volumetric data. The method generates uniformly triangulated, smooth meshes that accurately capture structural features. The process begins with the Ridge Field Transformation step that transforms the segmented input data, followed by the Mid-Polyline Extraction Algorithm that works on individual volume slices. Based on the connectivity of components, this step can result in either single or multiple polyline segments that represent the structural features. These segments form a coherent series, creating a backbone of regularly spaced points representing the mid-surface. Subsequently, we employ a Polyline Zipper Algorithm for triangulation that connects these polyline segments across neighboring slices, yielding a detailed triangulated mid-surface mesh. Results show that this method outperforms previous techniques in versatility, simplicity, and accuracy. Our approach is publicly available as a ParaView plugin at https://github.com/kaust-vislab/MidSurfer.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards softerware: Enabling personalization of interactive data representations for users with disabilities. 走向软件:为残疾用户实现交互式数据表示的个性化。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-12 DOI: 10.1109/MCG.2025.3609294
Frank Elavsky, Marita Vindedal, Ted Gies, Patrick Carrington, Dominik Moritz, Oystein Moseng

Accessible design for some may still produce barriers for others. This tension, called access friction, creates challenges for both designers and end-users with disabilities. To address this, we present the concept of softerware, a system design approach that provides end users with agency to meaningfully customize and adapt interfaces to their needs. To apply softerware to visualization, we assembled 195 data visualization customization options centered on the barriers we expect users with disabilities will experience. We built a prototype that applies a subset of these options and interviewed practitioners for feedback. Lastly, we conducted a design probe study with blind and low vision accessibility professionals to learn more about their challenges and visions for softerware. We observed access frictions between our participant's designs and they expressed that for softerware's success, current and future systems must be designed with accessible defaults, interoperability, persistence, and respect for a user's perceived effort-to-outcome ratio.

对某些人来说,无障碍设计可能仍然会给其他人带来障碍。这种紧张关系被称为访问摩擦,给设计师和残疾终端用户都带来了挑战。为了解决这个问题,我们提出了软件的概念,这是一种系统设计方法,为最终用户提供有意义的定制和调整接口以满足他们的需求的代理。为了将软件应用于可视化,我们集合了195个数据可视化定制选项,这些选项集中在残疾用户可能遇到的障碍上。我们构建了一个应用这些选项子集的原型,并采访了从业者以获得反馈。最后,我们对盲人和低视力无障碍专业人士进行了设计探索研究,以了解他们对软件的挑战和愿景。我们观察到参与者设计之间的访问摩擦,他们表示,为了软件的成功,当前和未来的系统必须具有可访问的默认值、互操作性、持久性,并尊重用户感知的努力与结果比。
{"title":"Towards softerware: Enabling personalization of interactive data representations for users with disabilities.","authors":"Frank Elavsky, Marita Vindedal, Ted Gies, Patrick Carrington, Dominik Moritz, Oystein Moseng","doi":"10.1109/MCG.2025.3609294","DOIUrl":"https://doi.org/10.1109/MCG.2025.3609294","url":null,"abstract":"<p><p>Accessible design for some may still produce barriers for others. This tension, called access friction, creates challenges for both designers and end-users with disabilities. To address this, we present the concept of softerware, a system design approach that provides end users with agency to meaningfully customize and adapt interfaces to their needs. To apply softerware to visualization, we assembled 195 data visualization customization options centered on the barriers we expect users with disabilities will experience. We built a prototype that applies a subset of these options and interviewed practitioners for feedback. Lastly, we conducted a design probe study with blind and low vision accessibility professionals to learn more about their challenges and visions for softerware. We observed access frictions between our participant's designs and they expressed that for softerware's success, current and future systems must be designed with accessible defaults, interoperability, persistence, and respect for a user's perceived effort-to-outcome ratio.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145056047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Virtual Staging of Indoor Panoramic Images via Multi-task Learning and Inverse Rendering. 基于多任务学习和反向渲染的室内全景图像虚拟舞台。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-03 DOI: 10.1109/MCG.2025.3605806
Uzair Shah, Sara Jashari, Muhammad Tukur, Mowafa Househ, Jens Schneider, Giovanni Pintore, Enrico Gobbetti, Marco Agus

Capturing indoor environments with 360° images provides a cost-effective method for creating immersive content. However, virtual staging - removing existing furniture and inserting new objects with realistic lighting - remains challenging. We present VISPI (Virtual Staging Pipeline for Single Indoor Panoramic Images), a framework that enables interactive restaging of indoor scenes from a single panoramic image. Our approach combines multi-task deep learning with real-time rendering to extract geometric, semantic, and material information from cluttered scenes. The system includes: i) a vision transformer that simultaneously predicts depth, normals, semantics, albedo, and material properties; ii) spherical Gaussian lighting estimation; iii) real-time editing for interactive object placement; iv) stereoscopic Multi-Center-Of-Projection generation for Head Mounted Display exploration. The framework processes input through two pathways: extracting clutter-free representations for virtual staging and estimating material properties including metallic and roughness signals. We evaluate VISPI on Structured3D and FutureHouse datasets, demonstrating applications in real estate visualization, interior design, and virtual environment creation.

用360°图像捕捉室内环境为创建沉浸式内容提供了一种经济有效的方法。然而,虚拟舞台-移除现有的家具并插入具有现实照明的新物体-仍然具有挑战性。我们提出了VISPI(单个室内全景图像的虚拟分级管道),这是一个框架,可以从单个全景图像中交互式地再现室内场景。我们的方法将多任务深度学习与实时渲染相结合,从混乱的场景中提取几何、语义和材料信息。该系统包括:i)同时预测深度、法线、语义、反照率和材料属性的视觉转换器;ii)球面高斯光照估计;Iii)交互式对象放置的实时编辑;iv)用于头戴式显示器探索的立体多中心投影生成。该框架通过两种途径处理输入:提取虚拟分期的无杂波表示和估计材料属性,包括金属和粗糙度信号。我们在Structured3D和FutureHouse数据集上评估了VISPI,展示了在房地产可视化、室内设计和虚拟环境创建方面的应用。
{"title":"Virtual Staging of Indoor Panoramic Images via Multi-task Learning and Inverse Rendering.","authors":"Uzair Shah, Sara Jashari, Muhammad Tukur, Mowafa Househ, Jens Schneider, Giovanni Pintore, Enrico Gobbetti, Marco Agus","doi":"10.1109/MCG.2025.3605806","DOIUrl":"https://doi.org/10.1109/MCG.2025.3605806","url":null,"abstract":"<p><p>Capturing indoor environments with 360° images provides a cost-effective method for creating immersive content. However, virtual staging - removing existing furniture and inserting new objects with realistic lighting - remains challenging. We present VISPI (Virtual Staging Pipeline for Single Indoor Panoramic Images), a framework that enables interactive restaging of indoor scenes from a single panoramic image. Our approach combines multi-task deep learning with real-time rendering to extract geometric, semantic, and material information from cluttered scenes. The system includes: i) a vision transformer that simultaneously predicts depth, normals, semantics, albedo, and material properties; ii) spherical Gaussian lighting estimation; iii) real-time editing for interactive object placement; iv) stereoscopic Multi-Center-Of-Projection generation for Head Mounted Display exploration. The framework processes input through two pathways: extracting clutter-free representations for virtual staging and estimating material properties including metallic and roughness signals. We evaluate VISPI on Structured3D and FutureHouse datasets, demonstrating applications in real estate visualization, interior design, and virtual environment creation.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144994462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tooth Completion and Reconstruction in Digital Orthodontics. 数字正畸的牙齿补全与重建。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-02 DOI: 10.1109/MCG.2025.3605266
Hao Yu, Longdu Liu, Shuangmin Chen, Shiqing Xin, Changhe Tu

In the field of digital orthodontics, dental models with complete roots are essential digital assets, particularly for visualization and treatment planning. However, intraoral scans typically capture only dental crowns, leaving roots missing. In this paper, we introduce a meticulously designed algorithmic pipeline to complete dental models while preserving crown geometry and mesh topology. Our pipeline begins with learning-based point cloud completion applied to existing dental crowns. We then reconstruct a complete tooth model, encompassing both the crown and root, to guide subsequent processing steps. Next, we restore the crown's original geometry and mesh topology using a strong Delaunay meshing structure; the correctness of this approach has been thoroughly established in existing literature. Finally, we optimize the transition region between crown and root using bi-harmonic smoothing. A key advantage of our approach is that the completed tooth model accurately maintains the geometry and mesh topology of the original crown, while also ensuring high-quality triangulation of dental roots.

在数字正畸领域,具有完整牙根的牙齿模型是必不可少的数字资产,特别是对于可视化和治疗计划。然而,口腔内扫描通常只捕获牙冠,而遗漏牙根。在本文中,我们介绍了一个精心设计的算法管道来完成牙齿模型,同时保留冠的几何形状和网格拓扑结构。我们的管道从基于学习的点云补全开始,应用于现有的牙冠。然后我们重建一个完整的牙齿模型,包括冠和根,以指导后续的处理步骤。接下来,我们使用强Delaunay网格结构恢复冠的原始几何形状和网格拓扑;这种方法的正确性已经在现有文献中得到了充分的证实。最后,我们利用双谐波平滑优化树冠和树根之间的过渡区域。我们的方法的一个关键优势是,完成的牙齿模型准确地保持了原始冠的几何形状和网格拓扑结构,同时也确保了高质量的牙根三角形。
{"title":"Tooth Completion and Reconstruction in Digital Orthodontics.","authors":"Hao Yu, Longdu Liu, Shuangmin Chen, Shiqing Xin, Changhe Tu","doi":"10.1109/MCG.2025.3605266","DOIUrl":"https://doi.org/10.1109/MCG.2025.3605266","url":null,"abstract":"<p><p>In the field of digital orthodontics, dental models with complete roots are essential digital assets, particularly for visualization and treatment planning. However, intraoral scans typically capture only dental crowns, leaving roots missing. In this paper, we introduce a meticulously designed algorithmic pipeline to complete dental models while preserving crown geometry and mesh topology. Our pipeline begins with learning-based point cloud completion applied to existing dental crowns. We then reconstruct a complete tooth model, encompassing both the crown and root, to guide subsequent processing steps. Next, we restore the crown's original geometry and mesh topology using a strong Delaunay meshing structure; the correctness of this approach has been thoroughly established in existing literature. Finally, we optimize the transition region between crown and root using bi-harmonic smoothing. A key advantage of our approach is that the completed tooth model accurately maintains the geometry and mesh topology of the original crown, while also ensuring high-quality triangulation of dental roots.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144979308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Cosmic View of Life on Earth: Hierarchical Visualization of Biological Data Using Astronomical Software. 地球生命的宇宙观:使用天文软件的生物数据分层可视化。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-01 DOI: 10.1109/MCG.2025.3591713
Wandrille Duchemin, Takanori Fujiwara, Hollister W Herhold, Elias Elmquist, David S Thaler, William Harcourt-Smith, Emma Broman, Alexander Bock, Brian P Abbott, Jacqueline K Faherty

A goal of data visualization is to advance the understanding of multiparameter, large-scale datasets. In astrophysics, scientists map celestial objects to understand the hierarchical structure of the universe. In biology, genetic sequences and biological characteristics uncover evolutionary relationships and patterns (e.g., variation within species and ecological associations). Our highly interdisciplinary project entitled "A Cosmic View of Life on Earth" adapts an immersive astrophysics visualization platform called OpenSpace to contextualize diverse biological data. Dimensionality reduction techniques harmonize biological information to create spatial representations in which data are interactively explored on flat screens and planetarium domes. Visualizations are enriched with geographic metadata, 3-D scans of specimens, and species-specific sonifications (e.g., bird songs). The "Cosmic View" project eases the dissemination of stories related to biological domains (e.g., insects, birds, mammals, and human migrations) and facilitates scientific discovery.

数据可视化的一个目标是促进对多参数、大规模数据集的理解。在天体物理学中,科学家绘制天体图以了解宇宙的层次结构。在生物学中,基因序列和生物学特征揭示了进化关系和模式(例如,物种内的变异和生态关联)。我们高度跨学科的项目名为“地球生命的宇宙观”,采用了一个名为OpenSpace的沉浸式天体物理学可视化平台,将各种生物数据置于背景中。降维技术协调生物信息,创建空间表示,其中数据在平面屏幕和天文馆圆顶上进行交互式探索。可视化通过地理元数据、标本的三维扫描和特定物种的声音(例如鸟鸣)来丰富。“宇宙观”项目简化了与生物领域(如昆虫、鸟类、哺乳动物、人类迁徙)有关的故事的传播,并促进了科学发现。
{"title":"A Cosmic View of Life on Earth: Hierarchical Visualization of Biological Data Using Astronomical Software.","authors":"Wandrille Duchemin, Takanori Fujiwara, Hollister W Herhold, Elias Elmquist, David S Thaler, William Harcourt-Smith, Emma Broman, Alexander Bock, Brian P Abbott, Jacqueline K Faherty","doi":"10.1109/MCG.2025.3591713","DOIUrl":"10.1109/MCG.2025.3591713","url":null,"abstract":"<p><p>A goal of data visualization is to advance the understanding of multiparameter, large-scale datasets. In astrophysics, scientists map celestial objects to understand the hierarchical structure of the universe. In biology, genetic sequences and biological characteristics uncover evolutionary relationships and patterns (e.g., variation within species and ecological associations). Our highly interdisciplinary project entitled \"A Cosmic View of Life on Earth\" adapts an immersive astrophysics visualization platform called OpenSpace to contextualize diverse biological data. Dimensionality reduction techniques harmonize biological information to create spatial representations in which data are interactively explored on flat screens and planetarium domes. Visualizations are enriched with geographic metadata, 3-D scans of specimens, and species-specific sonifications (e.g., bird songs). The \"Cosmic View\" project eases the dissemination of stories related to biological domains (e.g., insects, birds, mammals, and human migrations) and facilitates scientific discovery.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"93-106"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144692500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ProHap Explorer: Visualizing Haplotypes in Proteogenomic Datasets. ProHap Explorer:可视化蛋白质基因组数据集中的单倍型。
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-09-01 DOI: 10.1109/MCG.2025.3581736
Jakub Vasicek, Dafni Skiadopoulou, Ksenia G Kuznetsova, Lukas Kall, Marc Vaudel, Stefan Bruckner

In mass spectrometry-based proteomics, experts usually project data onto a single set of reference sequences, overlooking the influence of common haplotypes (combinations of genetic variants inherited together from a parent). We recently introduced ProHap, a tool for generating customized protein haplotype databases. Here, we present ProHap Explorer, a visualization interface designed to investigate the influence of common haplotypes on the human proteome. It enables users to explore haplotypes, their effects on protein sequences, and the identification of noncanonical peptides in public mass spectrometry datasets. The design builds on well-established representations in biological sequence analysis, ensuring familiarity for domain experts while integrating novel interactive elements tailored to proteogenomic data exploration. User interviews with proteomics experts confirmed the tool's utility, highlighting its ability to reveal whether haplotypes affect proteins of interest. By facilitating the intuitive exploration of proteogenomic variation, ProHap Explorer supports research in personalized medicine and the development of targeted therapies.

在以质谱为基础的蛋白质组学中,专家们通常将数据投射到一组参考序列上,忽略了共同单倍型(从父母那里遗传的基因变异的组合)的影响。我们最近介绍了ProHap,一个生成定制蛋白质单倍型数据库的工具。在这里,我们提出了ProHap Explorer,这是一个可视化界面,旨在研究常见单倍型对人类蛋白质组的影响。它使用户能够探索单倍型,它们对蛋白质序列的影响,以及在公共质谱数据集中鉴定非规范肽。该设计建立在生物序列分析中完善的表示的基础上,确保领域专家熟悉,同时集成为蛋白质基因组学数据探索量身定制的新型交互元素。用户与蛋白质组学专家的访谈证实了该工具的实用性,强调了其揭示单倍型是否影响感兴趣的蛋白质的能力。通过促进蛋白质基因组变异的直观探索,ProHap Explorer支持个性化医学研究和靶向治疗的开发。
{"title":"ProHap Explorer: Visualizing Haplotypes in Proteogenomic Datasets.","authors":"Jakub Vasicek, Dafni Skiadopoulou, Ksenia G Kuznetsova, Lukas Kall, Marc Vaudel, Stefan Bruckner","doi":"10.1109/MCG.2025.3581736","DOIUrl":"10.1109/MCG.2025.3581736","url":null,"abstract":"<p><p>In mass spectrometry-based proteomics, experts usually project data onto a single set of reference sequences, overlooking the influence of common haplotypes (combinations of genetic variants inherited together from a parent). We recently introduced ProHap, a tool for generating customized protein haplotype databases. Here, we present ProHap Explorer, a visualization interface designed to investigate the influence of common haplotypes on the human proteome. It enables users to explore haplotypes, their effects on protein sequences, and the identification of noncanonical peptides in public mass spectrometry datasets. The design builds on well-established representations in biological sequence analysis, ensuring familiarity for domain experts while integrating novel interactive elements tailored to proteogenomic data exploration. User interviews with proteomics experts confirmed the tool's utility, highlighting its ability to reveal whether haplotypes affect proteins of interest. By facilitating the intuitive exploration of proteogenomic variation, ProHap Explorer supports research in personalized medicine and the development of targeted therapies.</p>","PeriodicalId":55026,"journal":{"name":"IEEE Computer Graphics and Applications","volume":"PP ","pages":"64-77"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144337278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Computer Graphics and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1