首页 > 最新文献

Computers & Graphics-Uk最新文献

英文 中文
NEGS-Avatar: Normal Embedded Gaussians for 2D avatar from monocular video NEGS-Avatar:用于单目视频2D avatar的正常嵌入高斯分布
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-09 DOI: 10.1016/j.cag.2026.104538
Zedan Zheng , Yudi Tan , Zhuo Su , Fan Zhou , Baoquan Zhao
Creating realistic human avatars from monocular RGB videos is a long-standing and challenging problem. Existing implicit NeRF-based methods typically lack explicit geometric information in feature representation. Although 3D Gaussian Splatting (3DGS) has recently emerged as an explicit point-cloud-based alternative, information about geometric details like normal information is still missing in such an unstructured representation. In this paper, we present NEGS-Avatar, a novel approach to modeling animatable 2D human avatars from monocular videos using 3DGS. Our method incorporates normal information into 3D Gaussians as a learnable property to construct directed 3DGS to improve body appearance modeling. The normal information, along with other properties like positions, rotations and scales, is predicted based on the given body pose to model pose-dependent non-rigid deformation. The Gaussians are then transformed into actor posed space using linear blend skinning to realize pose animation. In addition, we develop a locality-aware adaptive density control strategy, which utilizes normal variance in local areas to facilitate effective Gaussain densification. Last but not the least, we propose to separate the specular and diffuse components for color prediction, thereby forming a more accurate, interpretable, and controllable appearance prediction model. Experimental results demonstrate that NEGS-Avatar achieves state-of-the-art performance both qualitatively and quantitatively, especially in the details of the clothing surface. The code is available at https://github.com/Zheng-ZD/NEGS-Avatar.git.
从单目RGB视频中创建逼真的人类化身是一个长期存在且具有挑战性的问题。现有基于隐式nerf的方法在特征表示中缺乏显式的几何信息。尽管3D高斯飞溅(3DGS)最近作为一种明确的基于点云的替代方案出现,但在这种非结构化表示中仍然缺少关于几何细节的信息,如正常信息。在本文中,我们提出了NEGS-Avatar,这是一种使用3DGS从单目视频中建模可动画的2D人类化身的新方法。我们的方法将法线信息作为一种可学习的属性整合到三维高斯模型中,以构建定向三维模型,以改善人体外观建模。法线信息,以及其他属性,如位置、旋转和尺度,都是基于给定的身体姿势来预测的,以模拟姿势相关的非刚性变形。然后利用线性混合蒙皮将高斯分布变换到演员的姿态空间,实现姿态动画。此外,我们开发了一种位置感知的自适应密度控制策略,该策略利用局部区域的正态方差来促进有效的高斯密度化。最后,我们提出将镜面和漫射分量分开进行颜色预测,从而形成一个更准确、可解释、可控的外观预测模型。实验结果表明,NEGS-Avatar在定性和定量上都达到了最先进的性能,特别是在服装表面的细节上。代码可在https://github.com/Zheng-ZD/NEGS-Avatar.git上获得。
{"title":"NEGS-Avatar: Normal Embedded Gaussians for 2D avatar from monocular video","authors":"Zedan Zheng ,&nbsp;Yudi Tan ,&nbsp;Zhuo Su ,&nbsp;Fan Zhou ,&nbsp;Baoquan Zhao","doi":"10.1016/j.cag.2026.104538","DOIUrl":"10.1016/j.cag.2026.104538","url":null,"abstract":"<div><div>Creating realistic human avatars from monocular RGB videos is a long-standing and challenging problem. Existing implicit NeRF-based methods typically lack explicit geometric information in feature representation. Although 3D Gaussian Splatting (3DGS) has recently emerged as an explicit point-cloud-based alternative, information about geometric details like normal information is still missing in such an unstructured representation. In this paper, we present NEGS-Avatar, a novel approach to modeling animatable 2D human avatars from monocular videos using 3DGS. Our method incorporates normal information into 3D Gaussians as a learnable property to construct directed 3DGS to improve body appearance modeling. The normal information, along with other properties like positions, rotations and scales, is predicted based on the given body pose to model pose-dependent non-rigid deformation. The Gaussians are then transformed into actor posed space using linear blend skinning to realize pose animation. In addition, we develop a locality-aware adaptive density control strategy, which utilizes normal variance in local areas to facilitate effective Gaussain densification. Last but not the least, we propose to separate the specular and diffuse components for color prediction, thereby forming a more accurate, interpretable, and controllable appearance prediction model. Experimental results demonstrate that NEGS-Avatar achieves state-of-the-art performance both qualitatively and quantitatively, especially in the details of the clothing surface. The code is available at <span><span>https://github.com/Zheng-ZD/NEGS-Avatar.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104538"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foreword to the Special Section on Smart Tools and Applications in Graphics (STAG 2024) 关于图形中的智能工具和应用的特别部分(STAG 2024)前言
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-01-23 DOI: 10.1016/j.cag.2026.104533
Andrea Giachetti, Umberto Castellani, Ariel Caputo, Valeria Garro, Nicola Capece
This Special Section contains extended and revised versions of selected papers presented at the 11th Conference on Smart Tools and Applications in Graphics (STAG 2024), held in Verona (Italy) on November 14–15, 2024. Three papers were selected by appointed members of the Program Committee; their extended versions were subsequently submitted and further reviewed by experts. The resulting collection comprises contributions spanning a broad range of topics, including navigation in mixed reality, reinforcement learning for intelligent agents in 3D environments, and interactive image relighting using neural networks.
本特别部分包含在2024年11月14日至15日在维罗纳(意大利)举行的第11届图形智能工具和应用会议(STAG 2024)上发表的选定论文的扩展和修订版本。三篇论文由项目委员会指定成员选出;随后提交了其扩展版本,并由专家进一步审查。由此产生的集合包括涵盖广泛主题的贡献,包括混合现实中的导航,3D环境中智能代理的强化学习以及使用神经网络的交互式图像重照明。
{"title":"Foreword to the Special Section on Smart Tools and Applications in Graphics (STAG 2024)","authors":"Andrea Giachetti,&nbsp;Umberto Castellani,&nbsp;Ariel Caputo,&nbsp;Valeria Garro,&nbsp;Nicola Capece","doi":"10.1016/j.cag.2026.104533","DOIUrl":"10.1016/j.cag.2026.104533","url":null,"abstract":"<div><div>This Special Section contains extended and revised versions of selected papers presented at the 11th Conference on Smart Tools and Applications in Graphics (STAG 2024), held in Verona (Italy) on November 14–15, 2024. Three papers were selected by appointed members of the Program Committee; their extended versions were subsequently submitted and further reviewed by experts. The resulting collection comprises contributions spanning a broad range of topics, including navigation in mixed reality, reinforcement learning for intelligent agents in 3D environments, and interactive image relighting using neural networks.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104533"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146080812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Golden anniversary of Computers & Graphics: A bibliometric overview 计算机与图形学的黄金周年纪念:文献计量学概述
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-05 DOI: 10.1016/j.cag.2026.104539
Muhammad Saqlain , José M. Merigó , Poom Kumam , Joaquim Jorge
Computers & Graphics celebrates its golden anniversary in 2025. Motivated by this special event, this study presents a comprehensive bibliometric analysis of the journal, identifying key research trends, frequently cited authors, institutions, countries, and major citation patterns. The work retrieves data from the Web of Science (WoS) core collection and Scopus databases and utilizes bibliometric tools such as VOS viewer and bibliometrix software. We analyse the keyword evolution; co-citation networks and bibliographic coupling of the documents published in Computers & Graphics. The distribution of topics indicates increased attention to artificial intelligence–based methods, including deep learning, point cloud processing, and virtual reality, alongside established rendering and simulation techniques. Additionally, the bibliometric analysis of productive authors, institutions and countries, indicate increased publication and citation activity associated with institutions in Asian countries, especially China. Beyond broader trends, this study also highlights Computers & Graphics’ recent initiatives that emphasize transparency and reproducibility, such as the graphics replicability stamp and the special sections, which bridge academic conferences and high-quality journal publications. This study serves as a reference for researchers seeking to understand the historical trajectory, emerging trends, and evolving editorial priorities in computer graphics research.
计算机图形公司将在2025年庆祝它的金婚纪念日。受这一特殊事件的影响,本研究对该期刊进行了全面的文献计量分析,确定了关键研究趋势、常被引作者、机构、国家和主要被引模式。该工作从Web of Science (WoS)核心馆藏和Scopus数据库中检索数据,并利用VOS viewer和bibliometrix软件等文献计量工具。我们分析了关键词的演变;《计算机图形学》上发表的文献的共引网络和书目耦合。主题的分布表明,人们越来越关注基于人工智能的方法,包括深度学习、点云处理和虚拟现实,以及已建立的渲染和模拟技术。此外,对高产作者、机构和国家的文献计量分析表明,亚洲国家(尤其是中国)与机构相关的出版物和引文活动增加。除了更广泛的趋势之外,这项研究还强调了Computers &; Graphics最近强调透明度和可重复性的举措,例如图形可复制印章和特殊部分,它们连接了学术会议和高质量的期刊出版物。本研究为研究人员寻求了解计算机图形学研究的历史轨迹、新兴趋势和不断发展的编辑优先事项提供了参考。
{"title":"Golden anniversary of Computers & Graphics: A bibliometric overview","authors":"Muhammad Saqlain ,&nbsp;José M. Merigó ,&nbsp;Poom Kumam ,&nbsp;Joaquim Jorge","doi":"10.1016/j.cag.2026.104539","DOIUrl":"10.1016/j.cag.2026.104539","url":null,"abstract":"<div><div><em>Computers &amp; Graphics</em> celebrates its golden anniversary in 2025. Motivated by this special event, this study presents a comprehensive bibliometric analysis of the journal, identifying key research trends, frequently cited authors, institutions, countries, and major citation patterns. The work retrieves data from the Web of Science (WoS) core collection and Scopus databases and utilizes bibliometric tools such as <em>VOS viewer</em> and <em>bibliometrix</em> software. We analyse the keyword evolution; co-citation networks and bibliographic coupling of the documents published in <em>Computers &amp; Graphics</em>. The distribution of topics indicates increased attention to artificial intelligence–based methods, including deep learning, point cloud processing, and virtual reality, alongside established rendering and simulation techniques. Additionally, the bibliometric analysis of productive authors, institutions and countries, indicate increased publication and citation activity associated with institutions in Asian countries, especially China. Beyond broader trends, this study also highlights <em>Computers &amp; Graphics’</em> recent initiatives that emphasize transparency and reproducibility, such as the graphics replicability stamp and the special sections, which bridge academic conferences and high-quality journal publications. This study serves as a reference for researchers seeking to understand the historical trajectory, emerging trends, and evolving editorial priorities in computer graphics research.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104539"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foreword to the Special Section on Shape Modeling International 2025 (SMI 2025) 形状建模国际2025 (SMI 2025)特别部分前言
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-11 DOI: 10.1016/j.cag.2026.104540
Hongwei Lin, Michela Mortara, Zichun Zhong
{"title":"Foreword to the Special Section on Shape Modeling International 2025 (SMI 2025)","authors":"Hongwei Lin,&nbsp;Michela Mortara,&nbsp;Zichun Zhong","doi":"10.1016/j.cag.2026.104540","DOIUrl":"10.1016/j.cag.2026.104540","url":null,"abstract":"","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104540"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sketch-guided stylized landscape cinemagraph synthesis 草图引导风格化景观电影合成
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-14 DOI: 10.1016/j.cag.2026.104547
Hao Jin , Hengyuan Chang , Xiaoxuan Xie , Zhengyang Wang , Xusheng Du , Shaojun Hu , Haoran Xie
Designing stylized cinemagraphs is challenging due to the difficulty in customizing complex and expressive flow elements. To achieve intuitive and detailed control of the generated cinemagraphs, sketches provide a feasible solution to convey personalized design requirements beyond text inputs. In this paper, we propose Sketch2Cinemagraph, a sketch-guided framework that enables the conditional generation of stylized cinemagraphs from freehand sketches. Sketch2Cinemagraph adopts text prompts for initial landscape generation and provides sketch controls for both spatial and motion cues. The latent diffusion model first generates target stylized landscape images along with realistic versions. Then, a pre-trained object detection model obtains masks for the flow regions. We propose a latent motion diffusion model to estimate motion field in fluid regions of the generated landscape images. The input motion sketches serve as the conditions to control the generated motion fields in the masked fluid regions with the prompt. To synthesize cinemagraph frames, the pixels within fluid regions are warped to target locations at each timestep using a U-Net based frame generator. The results verified that Sketch2Cinemagraph can generate aesthetically appealing stylized cinemagraphs with continuous temporal flow from sketch inputs. We showcase the advantages of Sketch2Cinemagraph through qualitative and quantitative comparisons against the state-of-the-art approaches.
设计风格化的电影动画是具有挑战性的,因为很难定制复杂和富有表现力的流元素。为了实现对生成的电影图形的直观和详细的控制,草图提供了一种可行的解决方案,可以在文本输入之外传达个性化的设计要求。在本文中,我们提出了Sketch2Cinemagraph,这是一个草图引导框架,可以从手绘草图中有条件地生成风格化的电影图。Sketch2Cinemagraph采用文本提示进行初始景观生成,并为空间和运动线索提供素描控制。潜在扩散模型首先生成目标风格化的景观图像以及现实版本。然后,预训练的目标检测模型获得流区域的掩码。我们提出了一个潜在的运动扩散模型来估计生成的景观图像的流体区域的运动场。输入的运动草图作为条件,以提示符控制被屏蔽流体区域中生成的运动场。为了合成电影帧,使用基于U-Net的帧生成器在每个时间步将流体区域内的像素扭曲到目标位置。结果验证了Sketch2Cinemagraph可以从草图输入中生成具有连续时间流的美学吸引力的风格化电影。我们通过对最先进的方法进行定性和定量比较,展示了Sketch2Cinemagraph的优势。
{"title":"Sketch-guided stylized landscape cinemagraph synthesis","authors":"Hao Jin ,&nbsp;Hengyuan Chang ,&nbsp;Xiaoxuan Xie ,&nbsp;Zhengyang Wang ,&nbsp;Xusheng Du ,&nbsp;Shaojun Hu ,&nbsp;Haoran Xie","doi":"10.1016/j.cag.2026.104547","DOIUrl":"10.1016/j.cag.2026.104547","url":null,"abstract":"<div><div>Designing stylized cinemagraphs is challenging due to the difficulty in customizing complex and expressive flow elements. To achieve intuitive and detailed control of the generated cinemagraphs, sketches provide a feasible solution to convey personalized design requirements beyond text inputs. In this paper, we propose Sketch2Cinemagraph, a sketch-guided framework that enables the conditional generation of stylized cinemagraphs from freehand sketches. Sketch2Cinemagraph adopts text prompts for initial landscape generation and provides sketch controls for both spatial and motion cues. The latent diffusion model first generates target stylized landscape images along with realistic versions. Then, a pre-trained object detection model obtains masks for the flow regions. We propose a latent motion diffusion model to estimate motion field in fluid regions of the generated landscape images. The input motion sketches serve as the conditions to control the generated motion fields in the masked fluid regions with the prompt. To synthesize cinemagraph frames, the pixels within fluid regions are warped to target locations at each timestep using a U-Net based frame generator. The results verified that Sketch2Cinemagraph can generate aesthetically appealing stylized cinemagraphs with continuous temporal flow from sketch inputs. We showcase the advantages of Sketch2Cinemagraph through qualitative and quantitative comparisons against the state-of-the-art approaches.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104547"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial Note for Issue 135 of Computers & Graphics 《计算机与图形学》第135期编辑说明
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-03-06 DOI: 10.1016/j.cag.2026.104567
Joaquim Jorge (Editor-in-Chief)
{"title":"Editorial Note for Issue 135 of Computers & Graphics","authors":"Joaquim Jorge (Editor-in-Chief)","doi":"10.1016/j.cag.2026.104567","DOIUrl":"10.1016/j.cag.2026.104567","url":null,"abstract":"","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104567"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Force-Scheme: A fast and accurate global dimensionality reduction method 增强力方案:一种快速、准确的全局降维方法
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-01-22 DOI: 10.1016/j.cag.2026.104536
Jaume Ros, Alessio Arleo, Fernando Paulovich
Global nonlinear Dimensionality Reduction (DR) methods excel at capturing complex features of datasets while preserving their overall high-dimensional structure when projecting them into a lower-dimensional space. Force-Scheme (FS) is one such method, used in a variety of domains. However, its use is still hindered by distortions and high computational cost. In this paper, we introduce Enhanced Force-Scheme (EFS), a revisited approach to solve the optimization problem posed by FS. We build on the core ideas of the original FS algorithm and introduce a more advanced optimization framework grounded in gradient-based optimization, which yields higher-quality layouts. Additionally, we elaborate on multiple strategies to accelerate the computation of projections using EFS, thereby facilitating its use on large datasets. Finally, we compare it with FS and other popular DR techniques and show that, among the methods tested, EFS best captures global structure while still performing well on local metrics.
全局非线性降维(DR)方法擅长捕获数据集的复杂特征,同时在将其投影到低维空间时保留其整体高维结构。力-方案(FS)就是这样一种方法,用于各种领域。然而,它的使用仍然受到扭曲和高计算成本的阻碍。在本文中,我们介绍了增强力方案(Enhanced Force-Scheme, EFS),这是一种重新设计的解决FS所带来的优化问题的方法。我们以原始FS算法的核心思想为基础,引入了基于梯度优化的更高级的优化框架,从而产生更高质量的布局。此外,我们详细阐述了使用EFS加速预测计算的多种策略,从而促进了其在大型数据集上的使用。最后,我们将其与FS和其他流行的DR技术进行了比较,并表明,在测试的方法中,EFS最好地捕获了全局结构,同时在局部指标上仍然表现良好。
{"title":"Enhanced Force-Scheme: A fast and accurate global dimensionality reduction method","authors":"Jaume Ros,&nbsp;Alessio Arleo,&nbsp;Fernando Paulovich","doi":"10.1016/j.cag.2026.104536","DOIUrl":"10.1016/j.cag.2026.104536","url":null,"abstract":"<div><div>Global nonlinear Dimensionality Reduction (DR) methods excel at capturing complex features of datasets while preserving their overall high-dimensional structure when projecting them into a lower-dimensional space. Force-Scheme (FS) is one such method, used in a variety of domains. However, its use is still hindered by distortions and high computational cost. In this paper, we introduce <em>Enhanced Force-Scheme</em> (EFS), a revisited approach to solve the optimization problem posed by FS. We build on the core ideas of the original FS algorithm and introduce a more advanced optimization framework grounded in gradient-based optimization, which yields higher-quality layouts. Additionally, we elaborate on multiple strategies to accelerate the computation of projections using EFS, thereby facilitating its use on large datasets. Finally, we compare it with FS and other popular DR techniques and show that, among the methods tested, EFS best captures global structure while still performing well on local metrics.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104536"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Systematic validation of LLM-generated structured data — A design space and remaining challenges 法学硕士生成的结构化数据的系统验证-一个设计空间和剩下的挑战
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-10 DOI: 10.1016/j.cag.2026.104545
Madhav Sachdeva , Christopher Narayanan , Marvin Wiedenkeller , Jana Sedlakova , Jürgen Bernard
Large language models (LLMs) are increasingly being used in academia and practice to generate structured data, supporting crucial data enrichment tasks such as imputing missing values, labeling data items, and generating synthetic datasets. However, these benefits rely on the validation of LLM-generated data to address known issues of LLMs, including hallucinations, inconsistencies, logical contradictions, and biases. Despite its importance and the significant growth of validation approaches in both diversity and count, the space opened up by these validation approaches remains unstructured. Based on a systematic literature review, we present a design space for approaches to the validation of LLM-generated structured data. The design space structures these approaches along two primary dimensions: Data Source and Granularity, and extends them with three complementary dimensions: Visualization techniques, Interaction techniques, and Workflow phases. Together, these dimensions form the descriptive, evaluative, and generative power of the design space. We apply the design space to demonstrate its utility through the analysis of three representative LLM-based validation approaches for structured data. Moreover, we reflect on the development process of Val-LLM, an interactive visual tool for multi-granularity validation, leveraging the design space as guideline in a novel approach. The results show that the design space enables researchers and practitioners to systematically characterize validation methods and guide the design of interactive systems for validation. We conclude by discussing limitations, remaining challenges, opportunities to extend the design space and to advance future validation research and practice.
大型语言模型(llm)在学术界和实践中越来越多地用于生成结构化数据,支持关键的数据充实任务,如输入缺失值、标记数据项和生成合成数据集。然而,这些好处依赖于llm生成的数据的验证,以解决llm的已知问题,包括幻觉、不一致、逻辑矛盾和偏见。尽管验证方法在多样性和数量上的重要性和显著增长,但这些验证方法所开辟的空间仍然是非结构化的。在系统的文献综述的基础上,我们提出了llm生成的结构化数据验证方法的设计空间。设计空间沿着两个主要维度构建这些方法:数据源和粒度,并通过三个补充维度扩展它们:可视化技术、交互技术和工作流阶段。这些维度共同构成了设计空间的描述性、评价性和生成能力。我们通过对结构化数据的三种代表性的基于llm的验证方法的分析,应用设计空间来展示其效用。此外,我们还反思了Val-LLM的开发过程,这是一种用于多粒度验证的交互式可视化工具,利用设计空间作为一种新方法的指导方针。结果表明,设计空间使研究人员和实践者能够系统地描述验证方法,并指导交互验证系统的设计。最后,我们讨论了扩展设计空间和推进未来验证研究和实践的局限性、仍然存在的挑战、机会。
{"title":"Systematic validation of LLM-generated structured data — A design space and remaining challenges","authors":"Madhav Sachdeva ,&nbsp;Christopher Narayanan ,&nbsp;Marvin Wiedenkeller ,&nbsp;Jana Sedlakova ,&nbsp;Jürgen Bernard","doi":"10.1016/j.cag.2026.104545","DOIUrl":"10.1016/j.cag.2026.104545","url":null,"abstract":"<div><div>Large language models (LLMs) are increasingly being used in academia and practice to generate structured data, supporting crucial data enrichment tasks such as imputing missing values, labeling data items, and generating synthetic datasets. However, these benefits rely on the validation of LLM-generated data to address known issues of LLMs, including hallucinations, inconsistencies, logical contradictions, and biases. Despite its importance and the significant growth of validation approaches in both diversity and count, the space opened up by these validation approaches remains unstructured. Based on a systematic literature review, we present a design space for approaches to the validation of LLM-generated structured data. The design space structures these approaches along two primary dimensions: <em>Data Source</em> and <em>Granularity</em>, and extends them with three complementary dimensions: <em>Visualization</em> techniques, Interaction techniques, and Workflow phases. Together, these dimensions form the descriptive, evaluative, and generative power of the design space. We apply the design space to demonstrate its utility through the analysis of three representative LLM-based validation approaches for structured data. Moreover, we reflect on the development process of <em>Val-LLM</em>, an interactive visual tool for multi-granularity validation, leveraging the design space as guideline in a novel approach. The results show that the design space enables researchers and practitioners to systematically characterize validation methods and guide the design of interactive systems for validation. We conclude by discussing limitations, remaining challenges, opportunities to extend the design space and to advance future validation research and practice.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104545"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foreword to special section on 15th Eurographics workshop on visual computing for biology and medicine 第15届欧洲图形学生物和医学视觉计算研讨会特别部分的前言
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-13 DOI: 10.1016/j.cag.2026.104546
Alessio Arleo , Jan Byška , Monique Meuschke
{"title":"Foreword to special section on 15th Eurographics workshop on visual computing for biology and medicine","authors":"Alessio Arleo ,&nbsp;Jan Byška ,&nbsp;Monique Meuschke","doi":"10.1016/j.cag.2026.104546","DOIUrl":"10.1016/j.cag.2026.104546","url":null,"abstract":"","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104546"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating LLMs’ abilities to create charts, a systematic approach 评估法学硕士创建图表的能力,这是一个系统的方法
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-04-01 Epub Date: 2026-02-18 DOI: 10.1016/j.cag.2026.104544
Maria Ribalta-Albado , Pere-Pau Vázquez
The use of generative models, especially those based on pretrained transformers, has become a common practice in code development. Tools such as GitHub Copilot, Cursor, and the direct use of conversational chatbots have proven useful to accelerate the development of applications. Unfortunately, generative models are unable to determine what is correct or wrong, and their outputs may contain errors. Their stochastic nature does not guarantee a single solution for the same problem, either. Furthermore, the output depends largely on the prompt issued by the user. To assess the capabilities of LLMs, some benchmarks have been proposed. Unfortunately, they often rely on ground truth data that may not be available. As a result, the extent to which modern LLMs can create charts needs further investigation. This work contributes to the understanding of the generative models’ ability to create charts in three ways: (a) Creating a dataset of prompts, data sources, and chart types to analyze, (b) Designing a set of systematic experiments that cover a wide range of commonly used charts, and variations of the visual variables, and (c) by empirically analyzing the performance of a large set of LLMs of different sizes, including Claude, CodeLlama, Gemini, Gemma, GPT4o, Llama 3.1, and Mixtral. Our results indicate that even the most advanced LLMs have room for improvement.
生成模型的使用,特别是那些基于预训练的变形器的模型,已经成为代码开发中的一种常见做法。事实证明,GitHub Copilot、Cursor和直接使用会话聊天机器人等工具对于加速应用程序的开发非常有用。不幸的是,生成模型无法确定什么是正确的或错误的,并且它们的输出可能包含错误。它们的随机特性也不能保证同一个问题有一个单一的解决方案。此外,输出在很大程度上取决于用户发出的提示。为了评估法学硕士的能力,已经提出了一些基准。不幸的是,他们往往依赖于可能无法获得的真实数据。因此,现代法学硕士在多大程度上可以创建图表需要进一步的研究。这项工作有助于理解生成模型以三种方式创建图表的能力:(a)创建提示、数据源和图表类型的数据集进行分析,(b)设计一组系统实验,涵盖广泛的常用图表和视觉变量的变化,以及(c)通过经验分析大量不同规模的llm的性能,包括Claude、CodeLlama、Gemini、Gemma、gpt40、Llama 3.1和Mixtral。我们的研究结果表明,即使是最先进的法学硕士也有改进的空间。
{"title":"Evaluating LLMs’ abilities to create charts, a systematic approach","authors":"Maria Ribalta-Albado ,&nbsp;Pere-Pau Vázquez","doi":"10.1016/j.cag.2026.104544","DOIUrl":"10.1016/j.cag.2026.104544","url":null,"abstract":"<div><div>The use of generative models, especially those based on pretrained transformers, has become a common practice in code development. Tools such as GitHub Copilot, Cursor, and the direct use of conversational chatbots have proven useful to accelerate the development of applications. Unfortunately, generative models are unable to determine what is correct or wrong, and their outputs may contain errors. Their stochastic nature does not guarantee a single solution for the same problem, either. Furthermore, the output depends largely on the prompt issued by the user. To assess the capabilities of LLMs, some benchmarks have been proposed. Unfortunately, they often rely on ground truth data that may not be available. As a result, the extent to which modern LLMs can create charts needs further investigation. This work contributes to the understanding of the generative models’ ability to create charts in three ways: <em>(a)</em> Creating a dataset of prompts, data sources, and chart types to analyze, <em>(b)</em> Designing a set of systematic experiments that cover a wide range of commonly used charts, and variations of the visual variables, and <em>(c)</em> by empirically analyzing the performance of a large set of LLMs of different sizes, including Claude, CodeLlama, Gemini, Gemma, GPT4o, Llama 3.1, and Mixtral. Our results indicate that even the most advanced LLMs have room for improvement.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"135 ","pages":"Article 104544"},"PeriodicalIF":2.8,"publicationDate":"2026-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147385464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers & Graphics-Uk
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1