首页 > 最新文献

Computers & Graphics-Uk最新文献

英文 中文
Imitation in relative terms using ReGAIL: Making motion controllers agile and transferable 使用ReGAIL进行相对模仿:使运动控制器变得灵活和可转移
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-10-20 DOI: 10.1016/j.cag.2025.104457
Paul Boursin , Yannis Kedadry , Tony Chevalier , Victor Zordan , Paul Kry , Sophie Grégoire , Marie-Paule Cani
We present an approach for training “agile” character control policies, able to produce a wide variety of motor skills from a single reference motion cycle. Our technique builds off of generative adversarial imitation learning (GAIL), with a key novelty of our approach being to provide modification to the observation map in order to improve agility and robustness. Namely, to support more agile behavior, we adjust the value measurements of the training discriminator through relative features - hence the name ReGAIL. Our state observations include both task relevant relative velocities and poses, as well as relative goal deviation information. In addition, to increase robustness of the resulting gaits, servo gains and damping values are included as part of the policy action to let the controller learn how to best combine tension and relaxation during motion. From a policy informed by a single reference motion, our resulting agent is able to maneuver as needed, at runtime, from walking forward to walking backward or sideways, turning and stepping nimbly. Moreover, thanks to the use of observations in relative frames, the trained controllers are robust to morphological changes of the simulated character, which makes adaptation to new morphologies straightforward. We demonstrate our approach for a humanoid and a quadruped, on both flat and sloped terrains, as well as provide ablation studies to validate the design choices of our framework. In addition, we present an application to prehistoric research, where being able to simulate hominids of specific morphologies on rough terrain is valuable with encouraging results.
我们提出了一种训练“敏捷”字符控制策略的方法,能够从单个参考运动周期中产生各种各样的运动技能。我们的技术建立在生成对抗模仿学习(GAIL)的基础上,我们方法的一个关键新颖之处在于对观察图进行修改,以提高敏捷性和鲁棒性。也就是说,为了支持更敏捷的行为,我们通过相关特征调整训练鉴别器的值度量——因此称为ReGAIL。我们的状态观测包括任务相关的相对速度和姿态,以及相对目标偏差信息。此外,为了增加所得步态的鲁棒性,伺服增益和阻尼值被包括作为策略动作的一部分,以使控制器学习如何在运动过程中最好地结合张力和松弛。根据由单个参考运动通知的策略,我们得到的代理能够在运行时根据需要进行机动,从向前走到向后或侧向走,灵活地转身和迈步。此外,由于使用了相对帧的观察,训练后的控制器对模拟特征的形态变化具有鲁棒性,这使得适应新的形态变得简单。我们展示了我们的方法,人形和四足动物,在平坦和倾斜的地形,并提供烧蚀研究,以验证我们的框架的设计选择。此外,我们提出了一个史前研究的应用,在那里能够模拟特定形态的原始人类在崎岖的地形是有价值的,令人鼓舞的结果。
{"title":"Imitation in relative terms using ReGAIL: Making motion controllers agile and transferable","authors":"Paul Boursin ,&nbsp;Yannis Kedadry ,&nbsp;Tony Chevalier ,&nbsp;Victor Zordan ,&nbsp;Paul Kry ,&nbsp;Sophie Grégoire ,&nbsp;Marie-Paule Cani","doi":"10.1016/j.cag.2025.104457","DOIUrl":"10.1016/j.cag.2025.104457","url":null,"abstract":"<div><div>We present an approach for training “agile” character control policies, able to produce a wide variety of motor skills from a single reference motion cycle. Our technique builds off of generative adversarial imitation learning (GAIL), with a key novelty of our approach being to provide modification to the observation map in order to improve agility and robustness. Namely, to support more agile behavior, we adjust the value measurements of the training discriminator through relative features - hence the name ReGAIL. Our state observations include both task relevant relative velocities and poses, as well as relative goal deviation information. In addition, to increase robustness of the resulting gaits, servo gains and damping values are included as part of the policy action to let the controller learn how to best combine tension and relaxation during motion. From a policy informed by a single reference motion, our resulting agent is able to maneuver as needed, at runtime, from walking forward to walking backward or sideways, turning and stepping nimbly. Moreover, thanks to the use of observations in relative frames, the trained controllers are robust to morphological changes of the simulated character, which makes adaptation to new morphologies straightforward. We demonstrate our approach for a humanoid and a quadruped, on both flat and sloped terrains, as well as provide ablation studies to validate the design choices of our framework. In addition, we present an application to prehistoric research, where being able to simulate hominids of specific morphologies on rough terrain is valuable with encouraging results.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104457"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature-driven compact representation model for analysis and visualization of large-scale multivariate SAMR data 面向大规模多变量SAMR数据分析与可视化的特征驱动紧凑表示模型
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-20 DOI: 10.1016/j.cag.2025.104331
Yang Yang , Yu Pei , Yi Cao
The storage overhead and I/O bottleneck of supercomputers creates a challenge in efficiently analyzing and visualizing large-scale multivariate SAMR data. It is thus necessary to greatly reduce the data size on the premise of maintaining data accuracy. In this paper, we propose a feature-driven compact representation model to handle structurally complex, high-dimensional, and nonlinear structured adaptive mesh refinement (SAMR) data for efficient storage, analysis, and visualization. We combine information-guided domain partition, distance-based dimensionality reduction, and error-bounded data representation to form a coherent three-component framework, achieving high compression ratios while ensuring low accuracy loss. Our approach addresses the key bottleneck in the visualization of large-scale multivariate SAMR data generated by massively parallel scientific simulations, namely the mutual restraint relationship between compression efficiency and data fidelity. We validate the effectiveness of our method using four datasets, the largest of which contains 4 billion grid points. Experimental results demonstrate that, compared with the state-of-the-art methods, our approach reduces data storage costs by approximately an order of magnitude while improving data reconstruction accuracy by nearly two orders of magnitude.
超级计算机的存储开销和I/O瓶颈给大规模多变量SAMR数据的有效分析和可视化带来了挑战。因此,有必要在保持数据准确性的前提下,大幅度减少数据的大小。在本文中,我们提出了一个特征驱动的紧凑表示模型来处理结构复杂、高维和非线性的结构化自适应网格细化(SAMR)数据,以实现高效的存储、分析和可视化。我们将信息引导的领域划分、基于距离的降维和错误边界的数据表示结合起来,形成了一个连贯的三组件框架,在保证低精度损失的同时实现了高压缩比。我们的方法解决了大规模并行科学模拟产生的大规模多元SAMR数据可视化的关键瓶颈,即压缩效率和数据保真度之间的相互约束关系。我们使用四个数据集验证了我们方法的有效性,其中最大的数据集包含40亿个网格点。实验结果表明,与最先进的方法相比,我们的方法将数据存储成本降低了大约一个数量级,同时将数据重建精度提高了近两个数量级。
{"title":"Feature-driven compact representation model for analysis and visualization of large-scale multivariate SAMR data","authors":"Yang Yang ,&nbsp;Yu Pei ,&nbsp;Yi Cao","doi":"10.1016/j.cag.2025.104331","DOIUrl":"10.1016/j.cag.2025.104331","url":null,"abstract":"<div><div>The storage overhead and I/O bottleneck of supercomputers creates a challenge in efficiently analyzing and visualizing large-scale multivariate SAMR data. It is thus necessary to greatly reduce the data size on the premise of maintaining data accuracy. In this paper, we propose a feature-driven compact representation model to handle structurally complex, high-dimensional, and nonlinear structured adaptive mesh refinement (SAMR) data for efficient storage, analysis, and visualization. We combine information-guided domain partition, distance-based dimensionality reduction, and error-bounded data representation to form a coherent three-component framework, achieving high compression ratios while ensuring low accuracy loss. Our approach addresses the key bottleneck in the visualization of large-scale multivariate SAMR data generated by massively parallel scientific simulations, namely the mutual restraint relationship between compression efficiency and data fidelity. We validate the effectiveness of our method using four datasets, the largest of which contains 4 billion grid points. Experimental results demonstrate that, compared with the state-of-the-art methods, our approach reduces data storage costs by approximately an order of magnitude while improving data reconstruction accuracy by nearly two orders of magnitude.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104331"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145110167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A unified framework for interactive visual graph matching via attribute-structure synchronization 基于属性-结构同步的交互式可视化图形匹配统一框架
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-15 DOI: 10.1016/j.cag.2025.104406
Yuhua Liu , Haoxuan Wang , Jiajia Kou , Ling Sun , Heyu Wang , Yongheng Wang , Yigang Wang , Jinchang Li , Zhiguang Zhou
In traditional graph retrieval tools, graph matching is commonly used to retrieve desired graphs from extensive graph datasets according to their structural similarities. However, in real applications, graph nodes have numerous attributes which also contain valuable information for evaluating similarities between graphs. Thus, to achieve superior graph matching results, it is crucial for graph retrieval tools to make full use of the attribute information in addition to structural information. We propose a novel framework for interactive visual graph matching. In the proposed framework, an attribute-structure synchronization method is developed for representing structural and attribute features in a unified embedding space based on Canonical Correlation Analysis (CCA). To support fast and interactive matching, our method provides users with intuitive visual query interfaces for traversing, filtering and searching for the target graph in the embedding space conveniently. With the designed interfaces, the users can also specify a new target graph with desired structural and semantic features. Besides, evaluation views are designed for easy validation and interpretation of the matching results. Case studies and quantitative comparisons on real-world datasets have demonstrated the superiorities of our proposed framework in graph matching and large graph exploration.
在传统的图检索工具中,图匹配通常是根据图的结构相似度从大量的图数据集中检索所需的图。然而,在实际应用中,图节点具有许多属性,这些属性还包含有价值的信息,用于评估图之间的相似性。因此,为了获得更好的图匹配结果,图检索工具除了充分利用结构信息外,还必须充分利用属性信息。我们提出了一种新的交互式视觉图形匹配框架。在该框架中,提出了一种基于典型相关分析(CCA)的属性-结构同步方法,用于在统一嵌入空间中表示结构和属性特征。为了支持快速和交互式匹配,我们的方法为用户提供了直观的可视化查询界面,方便用户在嵌入空间中遍历、过滤和搜索目标图。通过设计的接口,用户还可以指定具有所需结构和语义特征的新目标图。此外,还设计了评价视图,便于对匹配结果进行验证和解释。案例研究和对真实世界数据集的定量比较证明了我们提出的框架在图匹配和大图探索方面的优势。
{"title":"A unified framework for interactive visual graph matching via attribute-structure synchronization","authors":"Yuhua Liu ,&nbsp;Haoxuan Wang ,&nbsp;Jiajia Kou ,&nbsp;Ling Sun ,&nbsp;Heyu Wang ,&nbsp;Yongheng Wang ,&nbsp;Yigang Wang ,&nbsp;Jinchang Li ,&nbsp;Zhiguang Zhou","doi":"10.1016/j.cag.2025.104406","DOIUrl":"10.1016/j.cag.2025.104406","url":null,"abstract":"<div><div>In traditional graph retrieval tools, graph matching is commonly used to retrieve desired graphs from extensive graph datasets according to their structural similarities. However, in real applications, graph nodes have numerous attributes which also contain valuable information for evaluating similarities between graphs. Thus, to achieve superior graph matching results, it is crucial for graph retrieval tools to make full use of the attribute information in addition to structural information. We propose a novel framework for interactive visual graph matching. In the proposed framework, an attribute-structure synchronization method is developed for representing structural and attribute features in a unified embedding space based on Canonical Correlation Analysis (CCA). To support fast and interactive matching, our method provides users with intuitive visual query interfaces for traversing, filtering and searching for the target graph in the embedding space conveniently. With the designed interfaces, the users can also specify a new target graph with desired structural and semantic features. Besides, evaluation views are designed for easy validation and interpretation of the matching results. Case studies and quantitative comparisons on real-world datasets have demonstrated the superiorities of our proposed framework in graph matching and large graph exploration.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104406"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145120118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DeepSES: Learning solvent-excluded surfaces via neural signed distance fields DeepSES:通过神经符号距离场学习溶剂排除表面
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-23 DOI: 10.1016/j.cag.2025.104392
Niklas Merk, Anna Sterzik, Kai Lawonn
The solvent-excluded surface (SES) is essential for revealing molecular shape and solvent accessibility in applications such as molecular modeling, drug discovery, and protein folding. Its signed distance field (SDF) delivers a continuous, differentiable surface representation that enables efficient rendering, analysis, and interaction in volumetric visualization frameworks. However, analytic methods that compute the SDF of the SES cannot run at interactive rates on large biomolecular complexes, and grid-based methods tend to result in significant approximation errors, depending on molecular size and grid resolution. We address these limitations with DeepSES, a neural inference pipeline that predicts the SES SDF directly from the computationally simpler van der Waals (vdW) SDF on a fixed high-resolution grid. By employing an adaptive volume-filtering scheme that directs processing only to visible regions near the molecular surface, DeepSES yields interactive frame rates irrespective of molecule size. By offering multiple network configurations, DeepSES enables practitioners to balance inference time against prediction accuracy. In benchmarks on molecules ranging from one thousand to nearly four million atoms, our fastest configuration achieves real-time frame rates with a sub-angstrom mean error, while our highest-accuracy variant sustains interactive performance and outperforms state-of-the-art methods in terms of surface quality. By replacing costly algorithmic solvers with selective neural prediction, DeepSES provides a scalable, high-resolution solution for interactive biomolecular visualization.
溶剂排除表面(SES)在分子建模、药物发现和蛋白质折叠等应用中对于揭示分子形状和溶剂可及性至关重要。它的符号距离域(SDF)提供了一个连续的、可微的表面表示,在体积可视化框架中实现了高效的渲染、分析和交互。然而,计算SES的SDF的分析方法不能在大型生物分子复合物上以交互速率运行,并且基于网格的方法往往会导致显着的近似误差,这取决于分子大小和网格分辨率。我们使用DeepSES解决了这些限制,DeepSES是一种神经推理管道,可以直接从固定高分辨率网格上计算更简单的范德瓦尔斯(vdW) SDF预测SES SDF。通过采用自适应体积滤波方案,只对分子表面附近的可见区域进行处理,DeepSES产生的交互帧率与分子大小无关。通过提供多种网络配置,DeepSES使从业者能够平衡推理时间和预测准确性。在从1000到近400万个原子的分子基准测试中,我们最快的配置实现了亚埃平均误差的实时帧率,而我们最高精度的变体保持了交互性能,并在表面质量方面优于最先进的方法。通过用选择性神经预测取代昂贵的算法求解器,DeepSES为交互式生物分子可视化提供了可扩展的高分辨率解决方案。
{"title":"DeepSES: Learning solvent-excluded surfaces via neural signed distance fields","authors":"Niklas Merk,&nbsp;Anna Sterzik,&nbsp;Kai Lawonn","doi":"10.1016/j.cag.2025.104392","DOIUrl":"10.1016/j.cag.2025.104392","url":null,"abstract":"<div><div>The solvent-excluded surface (SES) is essential for revealing molecular shape and solvent accessibility in applications such as molecular modeling, drug discovery, and protein folding. Its signed distance field (SDF) delivers a continuous, differentiable surface representation that enables efficient rendering, analysis, and interaction in volumetric visualization frameworks. However, analytic methods that compute the SDF of the SES cannot run at interactive rates on large biomolecular complexes, and grid-based methods tend to result in significant approximation errors, depending on molecular size and grid resolution. We address these limitations with DeepSES, a neural inference pipeline that predicts the SES SDF directly from the computationally simpler van der Waals (vdW) SDF on a fixed high-resolution grid. By employing an adaptive volume-filtering scheme that directs processing only to visible regions near the molecular surface, DeepSES yields interactive frame rates irrespective of molecule size. By offering multiple network configurations, DeepSES enables practitioners to balance inference time against prediction accuracy. In benchmarks on molecules ranging from one thousand to nearly four million atoms, our fastest configuration achieves real-time frame rates with a sub-angstrom mean error, while our highest-accuracy variant sustains interactive performance and outperforms state-of-the-art methods in terms of surface quality. By replacing costly algorithmic solvers with selective neural prediction, DeepSES provides a scalable, high-resolution solution for interactive biomolecular visualization.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104392"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145160124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing agricultural remote sensing: A comprehensive review of deep supervised and Self-Supervised Learning for crop monitoring 推进农业遥感:作物监测的深度监督和自监督学习综述
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-26 DOI: 10.1016/j.cag.2025.104434
Mateus Pinto da Silva , Sabrina P.L.P. Correa , Mariana A.R. Schaefer , Julio C.S. Reis , Ian M. Nunes , Jefersson A. dos Santos , Hugo N. Oliveira
Deep Learning based on Remote Sensing has become a powerful tool to increase agricultural productivity, mitigate the effects of climate change, and monitor deforestation. However, there is a lack of standardization and appropriate taxonomic classification of the literature available in the context of informatics. Taking advantage of the categories already available in the literature, this paper provides an overview of the relevant literature categorized into five main applications: Parcel Segmentation, Crop Mapping, Crop Yielding, Land Use and Land Cover, and Change Detection. We review notable trends, including the transition from traditional to deep learning, convolutional models, recurrent and attention-based models, and generative strategies. We also map the use of Self-Supervised Learning through contrastive, non-contrastive, data masking and hybrid semi-supervised pretraining for the aforementioned applications with an experimental benchmark for Post-Harvest Crop Mapping models, and present our solution, SITS-Siam, which achieves top performance on two of the three datasets tested. In addition, we provide a comprehensive overview of publicly available datasets for these applications and also unlabeled datasets for Remote Sensing in general. We hope that our work can be useful as a guide for future work in this context. The benchmark code and the pre-trained weights are available in https://github.com/mateuspinto/rs-agriculture-survey-extended.
基于遥感的深度学习已成为提高农业生产力、减轻气候变化影响和监测森林砍伐的有力工具。然而,在信息学的背景下,文献缺乏标准化和适当的分类分类。利用现有文献的分类,本文概述了相关文献的五个主要应用:地块分割、作物制图、作物产量、土地利用和土地覆盖以及变化检测。我们回顾了一些值得注意的趋势,包括从传统学习到深度学习的转变,卷积模型,循环和基于注意力的模型,以及生成策略。我们还通过对比、非对比、数据屏蔽和混合半监督预训练对上述应用程序进行了映射,并对收获后作物映射模型进行了实验基准,并提出了我们的解决方案SITS-Siam,该解决方案在测试的三个数据集中的两个上达到了最佳性能。此外,我们还提供了这些应用的公开可用数据集的全面概述,以及一般遥感的未标记数据集。我们希望我们的工作能够对今后在这方面的工作起到指导作用。基准代码和预训练的权重可在https://github.com/mateuspinto/rs-agriculture-survey-extended中获得。
{"title":"Advancing agricultural remote sensing: A comprehensive review of deep supervised and Self-Supervised Learning for crop monitoring","authors":"Mateus Pinto da Silva ,&nbsp;Sabrina P.L.P. Correa ,&nbsp;Mariana A.R. Schaefer ,&nbsp;Julio C.S. Reis ,&nbsp;Ian M. Nunes ,&nbsp;Jefersson A. dos Santos ,&nbsp;Hugo N. Oliveira","doi":"10.1016/j.cag.2025.104434","DOIUrl":"10.1016/j.cag.2025.104434","url":null,"abstract":"<div><div>Deep Learning based on Remote Sensing has become a powerful tool to increase agricultural productivity, mitigate the effects of climate change, and monitor deforestation. However, there is a lack of standardization and appropriate taxonomic classification of the literature available in the context of informatics. Taking advantage of the categories already available in the literature, this paper provides an overview of the relevant literature categorized into five main applications: Parcel Segmentation, Crop Mapping, Crop Yielding, Land Use and Land Cover, and Change Detection. We review notable trends, including the transition from traditional to deep learning, convolutional models, recurrent and attention-based models, and generative strategies. We also map the use of Self-Supervised Learning through contrastive, non-contrastive, data masking and hybrid semi-supervised pretraining for the aforementioned applications with an experimental benchmark for Post-Harvest Crop Mapping models, and present our solution, SITS-Siam, which achieves top performance on two of the three datasets tested. In addition, we provide a comprehensive overview of publicly available datasets for these applications and also unlabeled datasets for Remote Sensing in general. We hope that our work can be useful as a guide for future work in this context. The benchmark code and the pre-trained weights are available in <span><span>https://github.com/mateuspinto/rs-agriculture-survey-extended</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104434"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SmartPoints: Enhanced local feature extraction and neighborhood diffusion network for 3D point cloud semantic segmentation SmartPoints:用于三维点云语义分割的增强局部特征提取和邻域扩散网络
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-11-11 DOI: 10.1016/j.cag.2025.104486
Ye Chen, Jian Lu, Jie Zhao, Xiaogai Chen, Kaibing Zhang
In recent years, transformer-based models have demonstrated strong performance in global information extraction. However, in 3D point cloud segmentation, such models still fall short when it comes to capturing local features and accurately identifying geometric and topological relationships. To address the resulting insufficiency in local feature extraction, we propose an enhanced local feature extraction and neighborhood diffusion network for 3D point cloud semantic segmentation (SmartPoints). First, our method aggregates local features from the input point set using a hierarchical feature fusion module (HFF), which enhances information interaction and dependency between different local regions. Next, the dual local topological structure perception module (DLTP) constructs two local topologies using positional and semantic information, respectively. An adaptive dynamic kernel is then designed to capture the mapping between the two local topologies, enhancing local feature representation. To address the challenge of unclear local neighborhood edge distinctions, which often lead to segmentation errors, we design a local neighborhood diffusion module (LND). This module achieves precise edge segmentation by enhancing target region features and suppressing non-target region features. Extensive experiments on benchmark datasets such as S3DIS, ScanNetV2 and SemanticKITTI demonstrate the superior segmentation performance of the proposed SmartPoints.
近年来,基于变压器的模型在全局信息提取中表现出了较强的性能。然而,在三维点云分割中,这种模型在捕捉局部特征和准确识别几何和拓扑关系方面仍然存在不足。为了解决局部特征提取的不足,我们提出了一种增强的局部特征提取和邻域扩散网络,用于3D点云语义分割(SmartPoints)。首先,该方法利用层次特征融合模块(HFF)从输入点集中聚合局部特征,增强了不同局部区域之间的信息交互和依赖关系;其次,双局部拓扑结构感知模块(dual local topology structure perception module, DLTP)分别利用位置信息和语义信息构建两个局部拓扑。然后设计了一个自适应动态核来捕获两个局部拓扑之间的映射,增强了局部特征表示。为了解决局部邻域边缘区分不清导致分割错误的问题,我们设计了一个局部邻域扩散模块(LND)。该模块通过增强目标区域特征和抑制非目标区域特征来实现精确的边缘分割。在S3DIS、ScanNetV2和SemanticKITTI等基准数据集上进行的大量实验证明了所提出的SmartPoints具有优越的分割性能。
{"title":"SmartPoints: Enhanced local feature extraction and neighborhood diffusion network for 3D point cloud semantic segmentation","authors":"Ye Chen,&nbsp;Jian Lu,&nbsp;Jie Zhao,&nbsp;Xiaogai Chen,&nbsp;Kaibing Zhang","doi":"10.1016/j.cag.2025.104486","DOIUrl":"10.1016/j.cag.2025.104486","url":null,"abstract":"<div><div>In recent years, transformer-based models have demonstrated strong performance in global information extraction. However, in 3D point cloud segmentation, such models still fall short when it comes to capturing local features and accurately identifying geometric and topological relationships. To address the resulting insufficiency in local feature extraction, we propose an enhanced local feature extraction and neighborhood diffusion network for 3D point cloud semantic segmentation (SmartPoints). First, our method aggregates local features from the input point set using a hierarchical feature fusion module (HFF), which enhances information interaction and dependency between different local regions. Next, the dual local topological structure perception module (DLTP) constructs two local topologies using positional and semantic information, respectively. An adaptive dynamic kernel is then designed to capture the mapping between the two local topologies, enhancing local feature representation. To address the challenge of unclear local neighborhood edge distinctions, which often lead to segmentation errors, we design a local neighborhood diffusion module (LND). This module achieves precise edge segmentation by enhancing target region features and suppressing non-target region features. Extensive experiments on benchmark datasets such as S3DIS, ScanNetV2 and SemanticKITTI demonstrate the superior segmentation performance of the proposed SmartPoints.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104486"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Preface to the special issue: SIBGRAPI 2024 tutorials 特刊前言:SIBGRAPI 2024教程
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-10-24 DOI: 10.1016/j.cag.2025.104460
Soraia Raupp Musse, Ricardo Marroquim, Zenilton K.G. Patrocínio
{"title":"Preface to the special issue: SIBGRAPI 2024 tutorials","authors":"Soraia Raupp Musse,&nbsp;Ricardo Marroquim,&nbsp;Zenilton K.G. Patrocínio","doi":"10.1016/j.cag.2025.104460","DOIUrl":"10.1016/j.cag.2025.104460","url":null,"abstract":"","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104460"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145417171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CeRF: Convolutional neural radiance derivative fields for new view synthesis CeRF:用于新视图合成的卷积神经辐射导数场
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-27 DOI: 10.1016/j.cag.2025.104447
Wenjie Liu, Ling You, Xiaoyan Yang, Dingbo Lu, Yang Li, Changbo Wang
Recently, Neural Radiance Fields (NeRF) has seen a surge in popularity, driven by its ability to generate high-fidelity novel view synthesized images. However, unexpected “floating ghost” artifacts usually emerge with limited training views and intricate optical phenomena. This issue stems from the inherent ambiguities in radiance fields, rooted in the fundamental volume rendering equation and the unrestricted learning paradigms in multi-layer perceptrons. In this paper, we introduce Convolutional Neural Radiance Fields (CeRF), a novel approach to model the derivatives of radiance along rays and solve the ambiguities through a fully neural rendering pipeline. To this end, a single-surface selection mechanism involving both a modified softmax function and an ideal point is proposed to implement our radiance derivative fields. Furthermore, a structured neural network architecture with 1D convolutional operations is employed to further boost the performance by extracting latent ray representations. Extensive experiments demonstrate the promising results of our proposed model compared with existing state-of-the-art approaches.
最近,神经辐射场(NeRF)因其能够生成高保真的新视图合成图像而受到广泛欢迎。然而,意想不到的“漂浮幽灵”人工制品通常出现在有限的训练视图和复杂的光学现象。这一问题源于辐射场固有的模糊性,根植于基本的体渲染方程和多层感知器的无限制学习范式。在本文中,我们介绍了卷积神经辐射场(CeRF),这是一种新的方法来模拟沿光线的辐射导数,并通过全神经渲染管道解决歧义。为此,提出了一种包含改进softmax函数和理想点的单曲面选择机制来实现我们的辐射导数场。此外,采用一维卷积操作的结构化神经网络架构,通过提取潜在射线表示进一步提高性能。与现有的最先进的方法相比,大量的实验证明了我们提出的模型的有希望的结果。
{"title":"CeRF: Convolutional neural radiance derivative fields for new view synthesis","authors":"Wenjie Liu,&nbsp;Ling You,&nbsp;Xiaoyan Yang,&nbsp;Dingbo Lu,&nbsp;Yang Li,&nbsp;Changbo Wang","doi":"10.1016/j.cag.2025.104447","DOIUrl":"10.1016/j.cag.2025.104447","url":null,"abstract":"<div><div>Recently, Neural Radiance Fields (NeRF) has seen a surge in popularity, driven by its ability to generate high-fidelity novel view synthesized images. However, unexpected “floating ghost” artifacts usually emerge with limited training views and intricate optical phenomena. This issue stems from the inherent ambiguities in radiance fields, rooted in the fundamental volume rendering equation and the unrestricted learning paradigms in multi-layer perceptrons. In this paper, we introduce Convolutional Neural Radiance Fields (CeRF), a novel approach to model the derivatives of radiance along rays and solve the ambiguities through a fully neural rendering pipeline. To this end, a single-surface selection mechanism involving both a modified softmax function and an ideal point is proposed to implement our radiance derivative fields. Furthermore, a structured neural network architecture with 1D convolutional operations is employed to further boost the performance by extracting latent ray representations. Extensive experiments demonstrate the promising results of our proposed model compared with existing state-of-the-art approaches.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104447"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145222446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Examining the attribution of gender and the perception of emotions in virtual humans 研究虚拟人类的性别归属和情感感知
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-09-29 DOI: 10.1016/j.cag.2025.104446
Victor Flávio de Andrade Araujo , Angelo Brandelli Costa , Soraia Raupp Musse
Virtual Humans (VHs) are becoming increasingly realistic, raising questions about how users perceive their gender and emotions. In this study, we investigate how textually assigned gender and visual facial features influence both gender attribution and emotion recognition in VHs. Two experiments were conducted. In the first, participants evaluated a nonbinary VH animated with expressions performed by both male and female actors. In the second part, participants assessed binary male and female VHs animated by either real actors or data-driven facial styles. Results show that users often rely on textual gender cues and facial features to assign gender to VHs. Emotion recognition was more accurate when expressions were performed by actresses or derived from facial styles, particularly in nonbinary models. Notably, participants more consistently attributed gender according to textual cues when the VH was visually androgynous, suggesting that the absence of strong gendered facial markers increases the reliance on textual information. These findings offer insights for designing more inclusive and perceptually coherent virtual agents.
虚拟人(VHs)正变得越来越逼真,这引发了用户如何感知自己的性别和情感的问题。在本研究中,我们探讨了文本分配的性别和视觉面部特征对视频性别归因和情绪识别的影响。进行了两个实验。在第一个实验中,参与者评估了一个由男女演员共同表演的非二元动画VH。在第二部分中,参与者评估由真人演员或数据驱动的面部风格制作的二元男性和女性VHs。结果表明,用户通常依靠文本性别线索和面部特征来确定录像带的性别。当表情由女演员表演或来自面部风格时,尤其是在非二元模型中,情感识别更准确。值得注意的是,当VH在视觉上是雌雄同体时,参与者更一致地根据文本线索来判断性别,这表明缺乏强烈的性别面部标记增加了对文本信息的依赖。这些发现为设计更具包容性和感知一致性的虚拟代理提供了见解。
{"title":"Examining the attribution of gender and the perception of emotions in virtual humans","authors":"Victor Flávio de Andrade Araujo ,&nbsp;Angelo Brandelli Costa ,&nbsp;Soraia Raupp Musse","doi":"10.1016/j.cag.2025.104446","DOIUrl":"10.1016/j.cag.2025.104446","url":null,"abstract":"<div><div>Virtual Humans (VHs) are becoming increasingly realistic, raising questions about how users perceive their gender and emotions. In this study, we investigate how textually assigned gender and visual facial features influence both gender attribution and emotion recognition in VHs. Two experiments were conducted. In the first, participants evaluated a nonbinary VH animated with expressions performed by both male and female actors. In the second part, participants assessed binary male and female VHs animated by either real actors or data-driven facial styles. Results show that users often rely on textual gender cues and facial features to assign gender to VHs. Emotion recognition was more accurate when expressions were performed by actresses or derived from facial styles, particularly in nonbinary models. Notably, participants more consistently attributed gender according to textual cues when the VH was visually androgynous, suggesting that the absence of strong gendered facial markers increases the reliance on textual information. These findings offer insights for designing more inclusive and perceptually coherent virtual agents.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104446"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145269537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MonoNeRF-DDP: Neural radiance fields from monocular endoscopic images with dense depth priors MonoNeRF-DDP:具有密集深度先验的单眼内窥镜图像的神经辐射场
IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-01 Epub Date: 2025-11-11 DOI: 10.1016/j.cag.2025.104487
Jinhua Liu , Dongjin Huang , Yongsheng Shi , Jiantao Qu
Synthesizing novel views from monocular endoscopic images is challenging due to sparse input views, occlusion of invalid regions, and soft tissue deformation. To tackle these challenges, we propose the neural radiance fields from monocular endoscopic images with dense depth priors, called MonoNeRF-DDP. The algorithm consists of two parts: preprocessing and normative depth-assisted reconstruction. In the preprocessing part, we use labelme to obtain mask images for invalid regions in endoscopy images, preventing their reconstruction. Then, to address the view sparsity problem, we fine-tuned a monocular depth estimation network to predict dense depth maps, enabling the recovery of scene depth information from sparse views during the neural radiance fields optimization process. In the normative depth-assisted reconstruction, to deal with the issues of soft tissue deformation and inaccurate depth information, we adopt neural radiance fields for dynamic scenes to take mask images and dense depth maps as additional inputs and utilize the proposed adaptive loss function to achieve self-supervised training. Experimental results show that MonoNeRF-DDP outperforms the best average values of competing algorithms across the real monocular endoscopic image dataset GastroSynth. MonoNeRF-DDP can reconstruct structurally accurate shapes, fine details, and highly realistic textures with only about 15 input images. Furthermore, a study of 14 medical-related participants indicates that MonoNeRF-DDP can more accurately observe the details of the disease sites and make more reliable preoperative diagnoses.
由于输入视图稀疏、无效区域遮挡和软组织变形,从单眼内窥镜图像合成新视图具有挑战性。为了解决这些挑战,我们提出了具有密集深度先验的单眼内窥镜图像的神经辐射场,称为MonoNeRF-DDP。该算法由预处理和规范深度辅助重建两部分组成。在预处理部分,我们使用标签对内窥镜图像中的无效区域进行掩码,防止其重构。然后,为了解决视图稀疏性问题,我们对单目深度估计网络进行了微调,以预测密集深度图,从而在神经辐射场优化过程中从稀疏视图中恢复场景深度信息。在规范的深度辅助重建中,为了解决软组织变形和深度信息不准确的问题,我们采用动态场景的神经辐射场,以掩模图像和密集深度图作为附加输入,利用提出的自适应损失函数实现自监督训练。实验结果表明,MonoNeRF-DDP在真实单眼内窥镜图像数据集GastroSynth上优于竞争算法的最佳平均值。MonoNeRF-DDP可以重建结构精确的形状,精细的细节,高度逼真的纹理,只有大约15个输入图像。此外,一项对14名医学相关参与者的研究表明,MonoNeRF-DDP可以更准确地观察疾病部位的细节,做出更可靠的术前诊断。
{"title":"MonoNeRF-DDP: Neural radiance fields from monocular endoscopic images with dense depth priors","authors":"Jinhua Liu ,&nbsp;Dongjin Huang ,&nbsp;Yongsheng Shi ,&nbsp;Jiantao Qu","doi":"10.1016/j.cag.2025.104487","DOIUrl":"10.1016/j.cag.2025.104487","url":null,"abstract":"<div><div>Synthesizing novel views from monocular endoscopic images is challenging due to sparse input views, occlusion of invalid regions, and soft tissue deformation. To tackle these challenges, we propose the neural radiance fields from monocular endoscopic images with dense depth priors, called MonoNeRF-DDP. The algorithm consists of two parts: preprocessing and normative depth-assisted reconstruction. In the preprocessing part, we use labelme to obtain mask images for invalid regions in endoscopy images, preventing their reconstruction. Then, to address the view sparsity problem, we fine-tuned a monocular depth estimation network to predict dense depth maps, enabling the recovery of scene depth information from sparse views during the neural radiance fields optimization process. In the normative depth-assisted reconstruction, to deal with the issues of soft tissue deformation and inaccurate depth information, we adopt neural radiance fields for dynamic scenes to take mask images and dense depth maps as additional inputs and utilize the proposed adaptive loss function to achieve self-supervised training. Experimental results show that MonoNeRF-DDP outperforms the best average values of competing algorithms across the real monocular endoscopic image dataset GastroSynth. MonoNeRF-DDP can reconstruct structurally accurate shapes, fine details, and highly realistic textures with only about 15 input images. Furthermore, a study of 14 medical-related participants indicates that MonoNeRF-DDP can more accurately observe the details of the disease sites and make more reliable preoperative diagnoses.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"133 ","pages":"Article 104487"},"PeriodicalIF":2.8,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers & Graphics-Uk
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1