首页 > 最新文献

Computers & Graphics-Uk最新文献

英文 中文
Imitation in relative terms using ReGAIL: Making motion controllers agile and transferable 使用ReGAIL进行相对模仿:使运动控制器变得灵活和可转移
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-20 DOI: 10.1016/j.cag.2025.104457
Paul Boursin , Yannis Kedadry , Tony Chevalier , Victor Zordan , Paul Kry , Sophie Grégoire , Marie-Paule Cani
We present an approach for training “agile” character control policies, able to produce a wide variety of motor skills from a single reference motion cycle. Our technique builds off of generative adversarial imitation learning (GAIL), with a key novelty of our approach being to provide modification to the observation map in order to improve agility and robustness. Namely, to support more agile behavior, we adjust the value measurements of the training discriminator through relative features - hence the name ReGAIL. Our state observations include both task relevant relative velocities and poses, as well as relative goal deviation information. In addition, to increase robustness of the resulting gaits, servo gains and damping values are included as part of the policy action to let the controller learn how to best combine tension and relaxation during motion. From a policy informed by a single reference motion, our resulting agent is able to maneuver as needed, at runtime, from walking forward to walking backward or sideways, turning and stepping nimbly. Moreover, thanks to the use of observations in relative frames, the trained controllers are robust to morphological changes of the simulated character, which makes adaptation to new morphologies straightforward. We demonstrate our approach for a humanoid and a quadruped, on both flat and sloped terrains, as well as provide ablation studies to validate the design choices of our framework. In addition, we present an application to prehistoric research, where being able to simulate hominids of specific morphologies on rough terrain is valuable with encouraging results.
我们提出了一种训练“敏捷”字符控制策略的方法,能够从单个参考运动周期中产生各种各样的运动技能。我们的技术建立在生成对抗模仿学习(GAIL)的基础上,我们方法的一个关键新颖之处在于对观察图进行修改,以提高敏捷性和鲁棒性。也就是说,为了支持更敏捷的行为,我们通过相关特征调整训练鉴别器的值度量——因此称为ReGAIL。我们的状态观测包括任务相关的相对速度和姿态,以及相对目标偏差信息。此外,为了增加所得步态的鲁棒性,伺服增益和阻尼值被包括作为策略动作的一部分,以使控制器学习如何在运动过程中最好地结合张力和松弛。根据由单个参考运动通知的策略,我们得到的代理能够在运行时根据需要进行机动,从向前走到向后或侧向走,灵活地转身和迈步。此外,由于使用了相对帧的观察,训练后的控制器对模拟特征的形态变化具有鲁棒性,这使得适应新的形态变得简单。我们展示了我们的方法,人形和四足动物,在平坦和倾斜的地形,并提供烧蚀研究,以验证我们的框架的设计选择。此外,我们提出了一个史前研究的应用,在那里能够模拟特定形态的原始人类在崎岖的地形是有价值的,令人鼓舞的结果。
{"title":"Imitation in relative terms using ReGAIL: Making motion controllers agile and transferable","authors":"Paul Boursin ,&nbsp;Yannis Kedadry ,&nbsp;Tony Chevalier ,&nbsp;Victor Zordan ,&nbsp;Paul Kry ,&nbsp;Sophie Grégoire ,&nbsp;Marie-Paule Cani","doi":"10.1016/j.cag.2025.104457","DOIUrl":"10.1016/j.cag.2025.104457","url":null,"abstract":"<div><div>We present an approach for training “agile” character control policies, able to produce a wide variety of motor skills from a single reference motion cycle. Our technique builds off of generative adversarial imitation learning (GAIL), with a key novelty of our approach being to provide modification to the observation map in order to improve agility and robustness. Namely, to support more agile behavior, we adjust the value measurements of the training discriminator through relative features - hence the name ReGAIL. Our state observations include both task relevant relative velocities and poses, as well as relative goal deviation information. In addition, to increase robustness of the resulting gaits, servo gains and damping values are included as part of the policy action to let the controller learn how to best combine tension and relaxation during motion. From a policy informed by a single reference motion, our resulting agent is able to maneuver as needed, at runtime, from walking forward to walking backward or sideways, turning and stepping nimbly. Moreover, thanks to the use of observations in relative frames, the trained controllers are robust to morphological changes of the simulated character, which makes adaptation to new morphologies straightforward. We demonstrate our approach for a humanoid and a quadruped, on both flat and sloped terrains, as well as provide ablation studies to validate the design choices of our framework. In addition, we present an application to prehistoric research, where being able to simulate hominids of specific morphologies on rough terrain is valuable with encouraging results.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104457"},"PeriodicalIF":2.8,"publicationDate":"2025-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic volumetric cloud modeling with edge details refinement 动态体积云建模与边缘细节细化
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-17 DOI: 10.1016/j.cag.2025.104456
Guanqun Wang , Shaoze Su , Bo Qin , Xinjie Wang
Real-time cloud modeling is crucial in computer graphics, atmospheric science, and virtual reality applications. Current methods for cloud modeling often inadequately represent realistic density variations and detailed cloud edges. To this end, we present a real-time modeling framework using multi-layer noise and coordinate-driven dynamics for realistic, detailed cloud modeling. The proposed framework utilizes 3D Perlin–Worley noise to construct volumetric cloud structures and incorporates density control to achieve layered visual effects. In addition, it employs a noise-driven flow map to refine edge details and enhances dynamic features such as cloud movement and deformation. To validate our approach, we conducted quantitative and qualitative experiments to compare our method with others and with real observational meteorological data, as well as a user study on perceptual quality. Results indicate that our method achieves better real-time performance and realism, particularly in edge details, which are essential for visual fidelity and interactive applications.
实时云建模在计算机图形学、大气科学和虚拟现实应用中是至关重要的。目前的云建模方法往往不能充分地表示真实的密度变化和详细的云边缘。为此,我们提出了一个使用多层噪声和坐标驱动动力学的实时建模框架,用于逼真、详细的云建模。该框架利用三维柏林-沃利噪声来构建体积云结构,并结合密度控制来实现分层的视觉效果。此外,它采用噪声驱动的流程图来细化边缘细节,并增强云运动和变形等动态特征。为了验证我们的方法,我们进行了定量和定性实验,将我们的方法与其他方法和实际观测气象数据进行比较,并对感知质量进行了用户研究。结果表明,我们的方法获得了更好的实时性和真实感,特别是在边缘细节方面,这对视觉保真和交互式应用至关重要。
{"title":"Dynamic volumetric cloud modeling with edge details refinement","authors":"Guanqun Wang ,&nbsp;Shaoze Su ,&nbsp;Bo Qin ,&nbsp;Xinjie Wang","doi":"10.1016/j.cag.2025.104456","DOIUrl":"10.1016/j.cag.2025.104456","url":null,"abstract":"<div><div>Real-time cloud modeling is crucial in computer graphics, atmospheric science, and virtual reality applications. Current methods for cloud modeling often inadequately represent realistic density variations and detailed cloud edges. To this end, we present a real-time modeling framework using multi-layer noise and coordinate-driven dynamics for realistic, detailed cloud modeling. The proposed framework utilizes 3D Perlin–Worley noise to construct volumetric cloud structures and incorporates density control to achieve layered visual effects. In addition, it employs a noise-driven flow map to refine edge details and enhances dynamic features such as cloud movement and deformation. To validate our approach, we conducted quantitative and qualitative experiments to compare our method with others and with real observational meteorological data, as well as a user study on perceptual quality. Results indicate that our method achieves better real-time performance and realism, particularly in edge details, which are essential for visual fidelity and interactive applications.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104456"},"PeriodicalIF":2.8,"publicationDate":"2025-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The design and evaluation of a tangible interface for cutaway visualization in Mixed Reality 混合现实中可视化界面的设计与评价
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-17 DOI: 10.1016/j.cag.2025.104454
Xuyu Li, Priyansh Jalan, John Dingliana
In this article, we report on research exploring tangible interfaces for authoring cutaway visualizations in Mixed Reality (MR). The proposed approach is to allow users to create the outline of a virtual cutaway section by physically tracing their fingers directly on the surface of a real object, to reveal virtual information within. Building on a prototype of such system, developed for the Microsoft HoloLens 2, we conducted user studies to compare the performance and feedback of participants using a tangible system in contrast to the completion of analogous tasks using conventional mid-air gestures. The experiments revealed that the tangible interface supported more accurate tracing, and participants consistently rated it as more intuitive and effective. These findings highlight the promise of tangible approaches and provide compelling direction for future work.
在这篇文章中,我们报告了在混合现实(MR)中探索创建可切割可视化的有形接口的研究。提出的方法是允许用户通过直接在真实物体表面上物理地追踪手指来创建虚拟剖面图的轮廓,以揭示其中的虚拟信息。基于为微软HoloLens 2开发的这种系统的原型,我们进行了用户研究,以比较参与者使用有形系统和使用传统半空中手势完成类似任务的表现和反馈。实验显示,有形界面支持更准确的追踪,参与者一致认为它更直观、更有效。这些发现突出了切实可行的方法的前景,并为未来的工作提供了令人信服的方向。
{"title":"The design and evaluation of a tangible interface for cutaway visualization in Mixed Reality","authors":"Xuyu Li,&nbsp;Priyansh Jalan,&nbsp;John Dingliana","doi":"10.1016/j.cag.2025.104454","DOIUrl":"10.1016/j.cag.2025.104454","url":null,"abstract":"<div><div>In this article, we report on research exploring tangible interfaces for authoring cutaway visualizations in Mixed Reality (MR). The proposed approach is to allow users to create the outline of a virtual cutaway section by physically tracing their fingers directly on the surface of a real object, to reveal virtual information within. Building on a prototype of such system, developed for the Microsoft HoloLens 2, we conducted user studies to compare the performance and feedback of participants using a tangible system in contrast to the completion of analogous tasks using conventional mid-air gestures. The experiments revealed that the tangible interface supported more accurate tracing, and participants consistently rated it as more intuitive and effective. These findings highlight the promise of tangible approaches and provide compelling direction for future work.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104454"},"PeriodicalIF":2.8,"publicationDate":"2025-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145417172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A personality-labeled semantic dataset from facial expressions, gaze, and head movement cues 来自面部表情、凝视和头部运动线索的个性标记语义数据集
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-16 DOI: 10.1016/j.cag.2025.104455
Srikar Kodavati, Wilhen Alberto Hui Mei, Anish Kanade, Funda Durupinar
Facial expressions, eye gaze, and head movements play key roles in personality perception and social interactions. However, datasets that capture these cues with personality labels are rare, which limits advancements in computational personality recognition and expressive virtual agents. To address this, we developed a compact, semantically interpretable dataset using temporal Action Units (AUs), gaze, and head movements labeled with Five-Factor personality traits. The dataset is annotated using an iterative strategy that employs the existing “First Impressions” dataset to train a deep-learning model for automated personality labeling. The model, which combines attention with convolutional components, is iteratively refined based on user feedback from small-scale user studies. Results show that combining AUs with gaze and head motion significantly improves personality trait prediction, outperforming state-of-the-art models in accuracy. We also analyze how AUs, gaze, and head parameters contribute to the perception of personality factors. The findings offer insights applicable to the animation of expressive virtual characters and behavioral sciences.
面部表情、眼神和头部运动在人格感知和社会互动中起着关键作用。然而,用人格标签捕获这些线索的数据集很少,这限制了计算人格识别和表达性虚拟代理的进步。为了解决这个问题,我们开发了一个紧凑的,语义可解释的数据集,使用时间动作单元(au),凝视和头部运动标记为五因素人格特征。数据集使用迭代策略进行注释,该策略使用现有的“第一印象”数据集来训练用于自动人格标记的深度学习模型。该模型结合了注意力和卷积分量,基于小规模用户研究的用户反馈进行迭代改进。结果表明,将AUs与凝视和头部运动相结合,显著提高了人格特质预测的准确性,优于目前最先进的模型。我们还分析了AUs、凝视和头部参数对人格因素感知的影响。这一发现为富有表现力的虚拟角色动画和行为科学提供了见解。
{"title":"A personality-labeled semantic dataset from facial expressions, gaze, and head movement cues","authors":"Srikar Kodavati,&nbsp;Wilhen Alberto Hui Mei,&nbsp;Anish Kanade,&nbsp;Funda Durupinar","doi":"10.1016/j.cag.2025.104455","DOIUrl":"10.1016/j.cag.2025.104455","url":null,"abstract":"<div><div>Facial expressions, eye gaze, and head movements play key roles in personality perception and social interactions. However, datasets that capture these cues with personality labels are rare, which limits advancements in computational personality recognition and expressive virtual agents. To address this, we developed a compact, semantically interpretable dataset using temporal Action Units (AUs), gaze, and head movements labeled with Five-Factor personality traits. The dataset is annotated using an iterative strategy that employs the existing “First Impressions” dataset to train a deep-learning model for automated personality labeling. The model, which combines attention with convolutional components, is iteratively refined based on user feedback from small-scale user studies. Results show that combining AUs with gaze and head motion significantly improves personality trait prediction, outperforming state-of-the-art models in accuracy. We also analyze how AUs, gaze, and head parameters contribute to the perception of personality factors. The findings offer insights applicable to the animation of expressive virtual characters and behavioral sciences.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104455"},"PeriodicalIF":2.8,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid visualisation of medical image: Surface and volume rendering 医学图像的混合可视化:表面和体绘制
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1016/j.cag.2025.104453
Belén Palma , Pablo Casanova-Salas , Jesús Gimeno , Sergio Casas-Yrurzum
Medical imaging plays a crucial role in diagnosis and surgical planning, but accurately merging segmented 3D models with volumetric medical data remains a significant challenge. Traditional surface-based 3D models provide a clear structural representation but may introduce segmentation inaccuracies, whereas Direct Volume Rendering (DVR) maintains all image details but lacks explicit surface definition. We present a novel hybrid rendering algorithm that seamlessly integrates Per-Pixel Linked Lists for accurate transparency sorting of 3D meshes with DVR techniques to ensure a smooth and precise blending of both data types. This approach captures the advantages of both methods—preserving the structural clarity of surface models while maintaining the depth and density information from the medical image. Our algorithm operates in real time and is compatible with virtual reality, enhancing spatial perception through stereoscopic visualisation. By allowing direct interaction with both segmented models and volumetric data, this technique reduces reliance on radiologists’ reports and improves surgical planning. Performance evaluations confirm real-time rendering, and an expert-based usability study further supports the method’s relevance for advanced medical visualisation applications.
医学成像在诊断和手术计划中起着至关重要的作用,但准确地将分割的3D模型与体积医学数据相结合仍然是一个重大挑战。传统的基于表面的3D模型提供了清晰的结构表示,但可能会引入不准确的分割,而直接体绘制(DVR)保留了所有图像细节,但缺乏明确的表面定义。我们提出了一种新的混合渲染算法,该算法无缝集成了逐像素链表,用于3D网格的精确透明排序和DVR技术,以确保两种数据类型的平滑和精确混合。该方法兼顾了两种方法的优点,既保留了表面模型的结构清晰度,又保留了医学图像的深度和密度信息。我们的算法实时运行,与虚拟现实兼容,通过立体可视化增强空间感知。通过允许与分割模型和体积数据直接交互,该技术减少了对放射科医生报告的依赖,并改善了手术计划。性能评估证实了实时渲染,一项基于专家的可用性研究进一步支持了该方法与高级医疗可视化应用的相关性。
{"title":"Hybrid visualisation of medical image: Surface and volume rendering","authors":"Belén Palma ,&nbsp;Pablo Casanova-Salas ,&nbsp;Jesús Gimeno ,&nbsp;Sergio Casas-Yrurzum","doi":"10.1016/j.cag.2025.104453","DOIUrl":"10.1016/j.cag.2025.104453","url":null,"abstract":"<div><div>Medical imaging plays a crucial role in diagnosis and surgical planning, but accurately merging segmented 3D models with volumetric medical data remains a significant challenge. Traditional surface-based 3D models provide a clear structural representation but may introduce segmentation inaccuracies, whereas Direct Volume Rendering (DVR) maintains all image details but lacks explicit surface definition. We present a novel hybrid rendering algorithm that seamlessly integrates Per-Pixel Linked Lists for accurate transparency sorting of 3D meshes with DVR techniques to ensure a smooth and precise blending of both data types. This approach captures the advantages of both methods—preserving the structural clarity of surface models while maintaining the depth and density information from the medical image. Our algorithm operates in real time and is compatible with virtual reality, enhancing spatial perception through stereoscopic visualisation. By allowing direct interaction with both segmented models and volumetric data, this technique reduces reliance on radiologists’ reports and improves surgical planning. Performance evaluations confirm real-time rendering, and an expert-based usability study further supports the method’s relevance for advanced medical visualisation applications.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104453"},"PeriodicalIF":2.8,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145321930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Designing and evaluating an immersive VR experience of a historic sailing ship in museum contexts 设计和评估博物馆背景下历史帆船的沉浸式VR体验
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-09 DOI: 10.1016/j.cag.2025.104439
Spyros Vosinakis , Panayiotis Koutsabasis , George Anastassakis , Andreas Papasalouros , Kostas Damianidis
Museums and exhibitions can benefit from immersive technologies by embodying visitors in rich interactive environments, where they can experience digitally reconstructed scenes and stories of the past. Nevertheless, public-space Virtual Reality (VR) interactions need to be short in duration, carefully designed to communicate the intended message, and optimized for the user experience, especially for first-time users. This paper contributes to the ongoing research on user experience in VR for cultural heritage through the presentation of the design and user evaluation of an installation that immerses users on board a historic sailing ship and has been part of a museum exhibition. We present the process of reconstructing the ship and developing the application with emphasis on design choices about the user experience (scene presentation, content delivery, navigation and interaction modes, assistance, etc.). We have performed a thorough user experience evaluation and present its results and our reflections on design issues regarding public VR installations for museums.
博物馆和展览可以受益于沉浸式技术,将参观者融入丰富的互动环境中,在那里他们可以体验到数字化重建的场景和过去的故事。然而,公共空间的虚拟现实(VR)交互需要持续时间短,精心设计以传达预期的信息,并针对用户体验进行优化,特别是对于首次用户。本文通过展示一个装置的设计和用户评价,为正在进行的VR文化遗产用户体验研究做出了贡献,该装置将用户沉浸在一艘历史悠久的帆船上,并已成为博物馆展览的一部分。我们介绍了重建船舶和开发应用程序的过程,重点是关于用户体验的设计选择(场景呈现,内容传递,导航和交互模式,辅助等)。我们进行了全面的用户体验评估,并介绍了其结果以及我们对博物馆公共VR装置设计问题的思考。
{"title":"Designing and evaluating an immersive VR experience of a historic sailing ship in museum contexts","authors":"Spyros Vosinakis ,&nbsp;Panayiotis Koutsabasis ,&nbsp;George Anastassakis ,&nbsp;Andreas Papasalouros ,&nbsp;Kostas Damianidis","doi":"10.1016/j.cag.2025.104439","DOIUrl":"10.1016/j.cag.2025.104439","url":null,"abstract":"<div><div>Museums and exhibitions can benefit from immersive technologies by embodying visitors in rich interactive environments, where they can experience digitally reconstructed scenes and stories of the past. Nevertheless, public-space Virtual Reality (VR) interactions need to be short in duration, carefully designed to communicate the intended message, and optimized for the user experience, especially for first-time users. This paper contributes to the ongoing research on user experience in VR for cultural heritage through the presentation of the design and user evaluation of an installation that immerses users on board a historic sailing ship and has been part of a museum exhibition. We present the process of reconstructing the ship and developing the application with emphasis on design choices about the user experience (scene presentation, content delivery, navigation and interaction modes, assistance, etc.). We have performed a thorough user experience evaluation and present its results and our reflections on design issues regarding public VR installations for museums.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104439"},"PeriodicalIF":2.8,"publicationDate":"2025-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Including reflections in real-time voxel-based global illumination 包括基于体素的实时全局照明的反射
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-06 DOI: 10.1016/j.cag.2025.104449
Alejandro Cosin-Ayerbe, Gustavo Patow
Despite advances in rendering techniques, achieving high-quality real-time global illumination remains a significant challenge in Computer Graphics. While offline methods produce photorealistic lighting effects by accurately simulating light transport, real-time approaches struggle with the computational complexity of global illumination, particularly when handling dynamic scenes and moving light sources. Existing solutions often rely on precomputed data structures or approximate techniques, which either lack flexibility or introduce artifacts that degrade visual fidelity. In this work, we build upon previous research on a voxel-based real-time global illumination method to efficiently incorporate reflections and interreflections for both static and dynamic objects. Our approach leverages a voxelized scene representation, combined with a strategy for ray tracing camera-visible reflections, to ensure accurate materials while maintaining high performance. Key contributions include: (i) a high-quality material system capable of diffuse, glossy, and specular interreflections for both static and dynamic scene objects (ii) a highly-performant screen-space material model with a low memory consumption; and (iii) an open-source full implementation for further research and development. Our method outperforms state-of-the-art academic and industrial techniques, achieving higher quality and better temporal stability without requiring excessive computational resources. By enabling real-time global illumination with reflections, our work lays the foundation for more advanced rendering systems, ultimately moving closer to the visual fidelity of offline rendering while maintaining interactivity.
尽管渲染技术取得了进步,但在计算机图形学中实现高质量的实时全局照明仍然是一个重大挑战。虽然离线方法通过精确模拟光传输产生逼真的照明效果,但实时方法与全局照明的计算复杂性作斗争,特别是在处理动态场景和移动光源时。现有的解决方案通常依赖于预先计算的数据结构或近似技术,它们要么缺乏灵活性,要么引入降低视觉保真度的工件。在这项工作中,我们建立在先前基于体素的实时全局照明方法的研究基础上,有效地结合静态和动态物体的反射和互反射。我们的方法利用体素化场景表示,结合光线追踪相机可见反射的策略,以确保准确的材料,同时保持高性能。主要贡献包括:(i)一个高质量的材料系统,能够对静态和动态场景对象进行漫反射、光滑和镜面互反射;(ii)一个高性能的屏幕空间材料模型,具有低内存消耗;(iii)为进一步的研究和开发提供一个开源的全面实现。我们的方法优于最先进的学术和工业技术,在不需要过多计算资源的情况下实现更高的质量和更好的时间稳定性。通过启用具有反射的实时全局照明,我们的工作为更高级的渲染系统奠定了基础,最终在保持交互性的同时更接近离线渲染的视觉保真度。
{"title":"Including reflections in real-time voxel-based global illumination","authors":"Alejandro Cosin-Ayerbe,&nbsp;Gustavo Patow","doi":"10.1016/j.cag.2025.104449","DOIUrl":"10.1016/j.cag.2025.104449","url":null,"abstract":"<div><div>Despite advances in rendering techniques, achieving high-quality real-time global illumination remains a significant challenge in Computer Graphics. While offline methods produce photorealistic lighting effects by accurately simulating light transport, real-time approaches struggle with the computational complexity of global illumination, particularly when handling dynamic scenes and moving light sources. Existing solutions often rely on precomputed data structures or approximate techniques, which either lack flexibility or introduce artifacts that degrade visual fidelity. In this work, we build upon previous research on a voxel-based real-time global illumination method to efficiently incorporate reflections and interreflections for both static and dynamic objects. Our approach leverages a voxelized scene representation, combined with a strategy for ray tracing camera-visible reflections, to ensure accurate materials while maintaining high performance. Key contributions include: (i) a high-quality material system capable of diffuse, glossy, and specular interreflections for both static and dynamic scene objects (ii) a highly-performant screen-space material model with a low memory consumption; and (iii) an open-source full implementation for further research and development. Our method outperforms state-of-the-art academic and industrial techniques, achieving higher quality and better temporal stability without requiring excessive computational resources. By enabling real-time global illumination with reflections, our work lays the foundation for more advanced rendering systems, ultimately moving closer to the visual fidelity of offline rendering while maintaining interactivity.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104449"},"PeriodicalIF":2.8,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PersonalityLens: Visualizing in-depth analysis for LLM-driven personality insights PersonalityLens:为法学硕士驱动的个性洞察提供可视化的深入分析
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-04 DOI: 10.1016/j.cag.2025.104452
Xiaoyi Wang , Jialong Ye , Guangtao Zhang , Honglei Guo
Large Language Models (LLMs) have demonstrated strong potential for text-based personality assessment and are increasingly adopted by domain experts as assistive tools. Rather than focusing on prediction accuracy, users now prioritize insight-driven analysis, using LLMs to explore large volumes of written and spoken language through simple verbal prompts. However, a gap remains between LLM-detected personality traits and users’ ability to contextualize these outputs within established psychological theories and mechanisms. Existing tools often lack support for multi-level insights and fail to capture the dynamic evolution of traits and facets over time, limiting deeper analysis. To address this, we propose PersonalityLens, a visual analysis tool designed to enhance insight discovery in personality analysis. Our design is informed by a comprehensive requirements analysis with domain experts and supports: (1) in-depth exploration of detected traits and their corresponding utterances, supporting insights at varying levels of granularity, (2) exploration of how personality traits and facets dynamically evolve in finer contexts over time, (3) alignment of traits and facets with psychological theories. We present two complementary case studies — one based on fictional TV dialogue and the other on therapeutic interactions — demonstrating PersonalityLens’s adaptability to diverse analytic goals and contexts. A qualitative think-aloud user study shows that PersonalityLens supports context-aware interpretation and insight discovery. Building on these findings, we outline design implications to inspire future research and enhance psychotherapy tools with integrated personality analysis for mental health support.
大型语言模型(llm)在基于文本的人格评估方面已经显示出强大的潜力,并且越来越多地被领域专家作为辅助工具采用。用户现在不再关注预测的准确性,而是优先考虑洞察力驱动的分析,使用llm通过简单的口头提示来探索大量的书面和口头语言。然而,llm检测到的人格特征与用户在既定心理学理论和机制中将这些输出情境化的能力之间仍然存在差距。现有的工具往往缺乏对多层次洞察的支持,无法捕捉特征和方面随时间的动态演变,从而限制了更深入的分析。为了解决这个问题,我们提出了PersonalityLens,这是一个视觉分析工具,旨在增强个性分析中的洞察力发现。我们的设计基于领域专家和支持的全面需求分析:(1)深入探索检测到的特征及其相应的话语,支持不同粒度水平的见解;(2)探索人格特征和方面如何随着时间的推移在更精细的环境中动态演变;(3)将特征和方面与心理学理论结合起来。我们提出了两个互补的案例研究——一个基于虚构的电视对话,另一个基于治疗互动——展示了PersonalityLens对不同分析目标和背景的适应性。一项定性的有声思考用户研究表明,PersonalityLens支持上下文感知解释和洞察力发现。在这些发现的基础上,我们概述了设计意义,以启发未来的研究,并通过综合人格分析来增强心理治疗工具,以提供心理健康支持。
{"title":"PersonalityLens: Visualizing in-depth analysis for LLM-driven personality insights","authors":"Xiaoyi Wang ,&nbsp;Jialong Ye ,&nbsp;Guangtao Zhang ,&nbsp;Honglei Guo","doi":"10.1016/j.cag.2025.104452","DOIUrl":"10.1016/j.cag.2025.104452","url":null,"abstract":"<div><div>Large Language Models (LLMs) have demonstrated strong potential for text-based personality assessment and are increasingly adopted by domain experts as assistive tools. Rather than focusing on prediction accuracy, users now prioritize insight-driven analysis, using LLMs to explore large volumes of written and spoken language through simple verbal prompts. However, a gap remains between LLM-detected personality traits and users’ ability to contextualize these outputs within established psychological theories and mechanisms. Existing tools often lack support for multi-level insights and fail to capture the dynamic evolution of traits and facets over time, limiting deeper analysis. To address this, we propose PersonalityLens, a visual analysis tool designed to enhance insight discovery in personality analysis. Our design is informed by a comprehensive requirements analysis with domain experts and supports: (1) in-depth exploration of detected traits and their corresponding utterances, supporting insights at varying levels of granularity, (2) exploration of how personality traits and facets dynamically evolve in finer contexts over time, (3) alignment of traits and facets with psychological theories. We present two complementary case studies — one based on fictional TV dialogue and the other on therapeutic interactions — demonstrating PersonalityLens’s adaptability to diverse analytic goals and contexts. A qualitative think-aloud user study shows that PersonalityLens supports context-aware interpretation and insight discovery. Building on these findings, we outline design implications to inspire future research and enhance psychotherapy tools with integrated personality analysis for mental health support.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104452"},"PeriodicalIF":2.8,"publicationDate":"2025-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CaRoLS: Condition-adaptive multi-level road layout synthesis CaRoLS:条件自适应多层次道路布局综合
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-03 DOI: 10.1016/j.cag.2025.104451
Tian Feng , Long Li , Weitao Li , Bo Li , Junao Shen
Synthesizing road layouts, which define the spatial structure of cities, is critical for many urban applications. Conventional deep learning methods, however, struggle to handle both unconditional and conditional inputs, and rarely capture the multi-level complexity of real road networks. We propose CaRoLS, a unified two-stage method for condition-adaptive multi-level road layout synthesis. Specifically, the Multi-level Layout Reconstruction stage uses a pre-trained variational autoencoder to encode a real-world road layout into a latent representation and then reconstructs the image. The Condition-adaptive Representation Generation stage employs a diffusion model to generate a latent representation from Gaussian noise, or from noise combined with an optional conditioning image containing natural and socio-economic information. This design balances computational efficiency with the ability to model continuous data. To further enhance output quality, we introduce a Condition-aware Decoder Block module that integrates global context and local details, replacing the standard U-Net decoder blocks in the diffusion model. Experiments on an Australian metropolitan dataset show that CaRoLS outperforms representative general and specialized synthesis methods. Compared to the current state-of-the-art methods, improvements reach up to 36.47% and 4.05% in image and topological metrics for the unconditional mode, and 56.25% and 3.18% in the conditional mode. These results demonstrate that CaRoLS generates multi-level road layouts with strong structural fidelity and high connectivity, and provides a unified pipeline for both unconditional and conditional synthesis.
综合道路布局定义了城市的空间结构,对许多城市应用至关重要。然而,传统的深度学习方法很难处理无条件和条件输入,并且很少捕捉到真实道路网络的多层次复杂性。提出了一种统一的两阶段方法CaRoLS,用于条件自适应多层次道路布局综合。具体来说,多层次布局重建阶段使用预训练的变分自编码器将真实世界的道路布局编码为潜在表示,然后重建图像。条件自适应表示生成阶段采用扩散模型从高斯噪声中生成潜在表示,或者从噪声与包含自然和社会经济信息的可选条件图像相结合中生成潜在表示。这种设计平衡了计算效率和对连续数据建模的能力。为了进一步提高输出质量,我们引入了一个条件感知解码器块模块,该模块集成了全局上下文和局部细节,取代了扩散模型中的标准U-Net解码器块。在澳大利亚大都市数据集上的实验表明,CaRoLS优于具有代表性的一般和专门合成方法。与目前最先进的方法相比,无条件模式的图像和拓扑指标的改进高达36.47%和4.05%,条件模式的改进高达56.25%和3.18%。这些结果表明,CaRoLS生成多层次的道路布局,具有较强的结构保真度和高连通性,并为无条件和条件综合提供了统一的管道。
{"title":"CaRoLS: Condition-adaptive multi-level road layout synthesis","authors":"Tian Feng ,&nbsp;Long Li ,&nbsp;Weitao Li ,&nbsp;Bo Li ,&nbsp;Junao Shen","doi":"10.1016/j.cag.2025.104451","DOIUrl":"10.1016/j.cag.2025.104451","url":null,"abstract":"<div><div>Synthesizing road layouts, which define the spatial structure of cities, is critical for many urban applications. Conventional deep learning methods, however, struggle to handle both unconditional and conditional inputs, and rarely capture the multi-level complexity of real road networks. We propose CaRoLS, a unified two-stage method for condition-adaptive multi-level road layout synthesis. Specifically, the Multi-level Layout Reconstruction stage uses a pre-trained variational autoencoder to encode a real-world road layout into a latent representation and then reconstructs the image. The Condition-adaptive Representation Generation stage employs a diffusion model to generate a latent representation from Gaussian noise, or from noise combined with an optional conditioning image containing natural and socio-economic information. This design balances computational efficiency with the ability to model continuous data. To further enhance output quality, we introduce a Condition-aware Decoder Block module that integrates global context and local details, replacing the standard U-Net decoder blocks in the diffusion model. Experiments on an Australian metropolitan dataset show that CaRoLS outperforms representative general and specialized synthesis methods. Compared to the current state-of-the-art methods, improvements reach up to 36.47% and 4.05% in image and topological metrics for the unconditional mode, and 56.25% and 3.18% in the conditional mode. These results demonstrate that CaRoLS generates multi-level road layouts with strong structural fidelity and high connectivity, and provides a unified pipeline for both unconditional and conditional synthesis.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104451"},"PeriodicalIF":2.8,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foreword to special section: Highlights from EuroVA 2024 特别部分前言:2024年欧锦赛亮点
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-01 DOI: 10.1016/j.cag.2025.104450
Hans-Jörg Schulz, Marco Angelini
{"title":"Foreword to special section: Highlights from EuroVA 2024","authors":"Hans-Jörg Schulz,&nbsp;Marco Angelini","doi":"10.1016/j.cag.2025.104450","DOIUrl":"10.1016/j.cag.2025.104450","url":null,"abstract":"","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104450"},"PeriodicalIF":2.8,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers & Graphics-Uk
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1