Pub Date : 2025-12-01Epub Date: 2025-11-11DOI: 10.1016/j.gmod.2025.101307
Michael Kofler, Michael Giritsch, Stefanie Elgeti
In this paper we present a lattice structure optimization approach by leveraging the capabilities of neural networks for implicit geometry representation. We employ the Deep Signed Distance Field (DeepSDF) method, where a continuous and low-dimensional latent space is introduced to encode the geometric information. In contrast to traditional topology optimization methods, this allows the restriction of the design space to specific geometries. In our case, the latent space is used to represent the geometry of different unit cells, that are stacked to form a lattice structure. Moreover, continuously varying the latent vector over the structure allows a functional grading and optimization. Unlike other lattice-structure optimization methods, we neither assume a large separation of scale nor periodicity. Instead, we perform a full-scale finite element analysis at each optimization step. The required mesh is obtained by a differentiable extension of the dual marching cubes algorithm, which enables gradient-based optimization.
在本文中,我们提出了一种利用神经网络的能力进行隐式几何表示的晶格结构优化方法。我们采用深度签名距离场(Deep Signed Distance Field, DeepSDF)方法,在该方法中引入一个连续的低维潜在空间来编码几何信息。与传统的拓扑优化方法相比,这种方法允许将设计空间限制为特定的几何形状。在我们的案例中,潜在空间用于表示不同单元格的几何形状,这些单元格堆叠形成晶格结构。此外,在结构上连续变化潜在向量允许功能分级和优化。与其他晶格结构优化方法不同,我们既没有假设大的尺度分离,也没有假设周期性。相反,我们在每个优化步骤中执行全面的有限元分析。通过对偶行进立方体算法的可微扩展获得所需的网格,从而实现基于梯度的优化。
{"title":"Structural optimization of lattice structures using deep neural networks as geometry representation","authors":"Michael Kofler, Michael Giritsch, Stefanie Elgeti","doi":"10.1016/j.gmod.2025.101307","DOIUrl":"10.1016/j.gmod.2025.101307","url":null,"abstract":"<div><div>In this paper we present a lattice structure optimization approach by leveraging the capabilities of neural networks for implicit geometry representation. We employ the Deep Signed Distance Field (DeepSDF) method, where a continuous and low-dimensional latent space is introduced to encode the geometric information. In contrast to traditional topology optimization methods, this allows the restriction of the design space to specific geometries. In our case, the latent space is used to represent the geometry of different unit cells, that are stacked to form a lattice structure. Moreover, continuously varying the latent vector over the structure allows a functional grading and optimization. Unlike other lattice-structure optimization methods, we neither assume a large separation of scale nor periodicity. Instead, we perform a full-scale finite element analysis at each optimization step. The required mesh is obtained by a differentiable extension of the dual marching cubes algorithm, which enables gradient-based optimization.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"142 ","pages":"Article 101307"},"PeriodicalIF":2.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145528885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-11-05DOI: 10.1016/j.gmod.2025.101303
Zhan Wang , Junhao Wang , Zongpu Li , Hao Su , Pei Lv , Mingliang Xu
AI-aided floorplan design is a longstanding task in computer graphics. However, most of the existing methods focus on generating floorplans by limited architecture-level elements (e.g., room sizes, positions, and adjacencies), which ignore environmental factors and do not support customized designs. In this paper, we propose FlexPlan, an interactive approach for high-flexibility floorplan design. In FlexPlan, we propose a novel graph structure, named ArchiGraph, which enables flexible editing more comprehensive layout elements (e.g., architectures, environments, human needs) in a floorplan. First, we match similar floorplans according to the input architecture and environment features. Then, leveraging ArchiGraph, we interactively produce rooms’ attributes and quickly output the vectorized floorplans. For ArchiGraph, we design a RelationNet to predict room adjacencies, and propose a BoxNet to generate high-quality room boxes. Subjective and objective experiments show that our method is compatible with generating diverse complex floorplans (e.g., floorplans with irregular layout boundaries and room shapes). Compared with the state-of-the-art methods, our method can produce higher quality floorplans, and increase the speed of layout generation by nearly 20 times at most.
{"title":"FlexPlan: High-flexibility interactive floorplan design based on ArchiGraph","authors":"Zhan Wang , Junhao Wang , Zongpu Li , Hao Su , Pei Lv , Mingliang Xu","doi":"10.1016/j.gmod.2025.101303","DOIUrl":"10.1016/j.gmod.2025.101303","url":null,"abstract":"<div><div>AI-aided floorplan design is a longstanding task in computer graphics. However, most of the existing methods focus on generating floorplans by limited architecture-level elements (e.g., room sizes, positions, and adjacencies), which ignore environmental factors and do not support customized designs. In this paper, we propose FlexPlan, an interactive approach for high-flexibility floorplan design. In FlexPlan, we propose a novel graph structure, named ArchiGraph, which enables flexible editing more comprehensive layout elements (e.g., architectures, environments, human needs) in a floorplan. First, we match similar floorplans according to the input architecture and environment features. Then, leveraging ArchiGraph, we interactively produce rooms’ attributes and quickly output the vectorized floorplans. For ArchiGraph, we design a RelationNet to predict room adjacencies, and propose a BoxNet to generate high-quality room boxes. Subjective and objective experiments show that our method is compatible with generating diverse complex floorplans (e.g., floorplans with irregular layout boundaries and room shapes). Compared with the state-of-the-art methods, our method can produce higher quality floorplans, and increase the speed of layout generation by nearly 20 times at most.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"142 ","pages":"Article 101303"},"PeriodicalIF":2.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145474232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-13DOI: 10.1016/j.gmod.2025.101296
Shuxian Cai , Juan Cao , Bailin Deng , Zhonggui Chen
Sharp feature lines provide critical structural information in 3D models and are essential for geometric processing. However, the performance of existing algorithms for extracting feature lines from point clouds remains sensitive to the quality of the input data. This paper introduces an algorithm specifically designed to extract feature lines from 3D point clouds. The algorithm calculates the winding number for each point and uses variations in this number within edge regions to identify feature points. These feature points are then mapped onto a cuboid structure to obtain key feature points and capture neighboring relationships. Finally, feature lines are fitted based on the connectivity of key feature points. Extensive experiments demonstrate that this algorithm not only accurately detects feature points on potential sharp edges, but also outperforms existing methods in extracting subtle feature lines and handling complex point clouds.
{"title":"Feature line extraction based on winding number","authors":"Shuxian Cai , Juan Cao , Bailin Deng , Zhonggui Chen","doi":"10.1016/j.gmod.2025.101296","DOIUrl":"10.1016/j.gmod.2025.101296","url":null,"abstract":"<div><div>Sharp feature lines provide critical structural information in 3D models and are essential for geometric processing. However, the performance of existing algorithms for extracting feature lines from point clouds remains sensitive to the quality of the input data. This paper introduces an algorithm specifically designed to extract feature lines from 3D point clouds. The algorithm calculates the winding number for each point and uses variations in this number within edge regions to identify feature points. These feature points are then mapped onto a cuboid structure to obtain key feature points and capture neighboring relationships. Finally, feature lines are fitted based on the connectivity of key feature points. Extensive experiments demonstrate that this algorithm not only accurately detects feature points on potential sharp edges, but also outperforms existing methods in extracting subtle feature lines and handling complex point clouds.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101296"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144829844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-15DOI: 10.1016/j.gmod.2025.101297
Hao Yu , Longdu Liu , Shuangmin Chen , Lin Lu , Yuanfeng Zhou , Shiqing Xin , Changhe Tu
The rapid evolution of digital orthodontics has highlighted a critical need for automated treatment planning systems that balance computational efficiency with clinical reliability. However, existing methods still suffer from several limitations, including excessive clinician involvement (accounting for over 35% of treatment planning time), reliance on empirically defined key frames, and limited biomechanical plausibility, particularly in cases of severe dental crowding. This paper proposes a novel collision-free optimization framework to address these issues simultaneously. Our method defines a total movement energy function evaluated over each tooth’s pose at intermediate time frames. This energy is minimized iteratively using a steepest descent strategy. A rollback mechanism is employed: if inter-tooth penetration is detected during an update, the step size is halved repeatedly until collisions are eliminated. The framework allows flexible control over the number of intermediate frames to enforce a strict constraint on per-tooth displacement, limiting it to 0.2 mm translation or rotation every 10 to 14 days. Clinical evaluations show that the proposed algorithm can generate desirable and clinically valid tooth movement plans, even in complex cases, while significantly reducing the need for manual intervention.
{"title":"Collision-free path planning method for digital orthodontic treatment","authors":"Hao Yu , Longdu Liu , Shuangmin Chen , Lin Lu , Yuanfeng Zhou , Shiqing Xin , Changhe Tu","doi":"10.1016/j.gmod.2025.101297","DOIUrl":"10.1016/j.gmod.2025.101297","url":null,"abstract":"<div><div>The rapid evolution of digital orthodontics has highlighted a critical need for automated treatment planning systems that balance computational efficiency with clinical reliability. However, existing methods still suffer from several limitations, including excessive clinician involvement (accounting for over 35% of treatment planning time), reliance on empirically defined key frames, and limited biomechanical plausibility, particularly in cases of severe dental crowding. This paper proposes a novel collision-free optimization framework to address these issues simultaneously. Our method defines a total movement energy function evaluated over each tooth’s pose at intermediate time frames. This energy is minimized iteratively using a steepest descent strategy. A rollback mechanism is employed: if inter-tooth penetration is detected during an update, the step size is halved repeatedly until collisions are eliminated. The framework allows flexible control over the number of intermediate frames to enforce a strict constraint on per-tooth displacement, limiting it to 0.2 mm translation or <span><math><mrow><mn>2</mn><mo>°</mo></mrow></math></span> rotation every 10 to 14 days. Clinical evaluations show that the proposed algorithm can generate desirable and clinically valid tooth movement plans, even in complex cases, while significantly reducing the need for manual intervention.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101297"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144842157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-05DOI: 10.1016/j.gmod.2025.101288
Fangfang Zhou , Haiyu Shen , Mingzhen Li , Ying Zhao , Chongke Bi
Porous materials (e.g., weathered stone, industrial coatings) exhibit complex optical effects due to their micrograin and pore structures, posing challenges for photorealistic rendering. Explicit geometry models struggle to characterize their micrograin distributions at microscopic scales, while single-scattering microfacet model fails to accurately capture the multiple-scattering effects and causes energy non-conservation artifacts, manifesting as unrealistic luminance decay. We propose an enhanced micrograin BSDF model that accurately accounts for multiple scattering. First, we introduce a visible normal distribution function (VNDF) sampling method via rejection sampling. Building on VNDF sampling, we derive a position-free microsurface formulation incorporating both inter-micrograin and micrograin-to-base interactions. Furthermore, we propose a practical random walk method to simulate microsurface scattering, which accurately solves the derived formulation. Our micrograin BSDF model effectively eliminates the energy loss artifacts inherent in the previous model while significantly reducing noise, providing a physically accurate yet artistically controllable solution for rendering porous materials.
{"title":"Position-free multiple-scattering computations for micrograin BSDF model","authors":"Fangfang Zhou , Haiyu Shen , Mingzhen Li , Ying Zhao , Chongke Bi","doi":"10.1016/j.gmod.2025.101288","DOIUrl":"10.1016/j.gmod.2025.101288","url":null,"abstract":"<div><div>Porous materials (e.g., weathered stone, industrial coatings) exhibit complex optical effects due to their micrograin and pore structures, posing challenges for photorealistic rendering. Explicit geometry models struggle to characterize their micrograin distributions at microscopic scales, while single-scattering microfacet model fails to accurately capture the multiple-scattering effects and causes energy non-conservation artifacts, manifesting as unrealistic luminance decay. We propose an enhanced micrograin BSDF model that accurately accounts for multiple scattering. First, we introduce a visible normal distribution function (VNDF) sampling method via rejection sampling. Building on VNDF sampling, we derive a position-free microsurface formulation incorporating both inter-micrograin and micrograin-to-base interactions. Furthermore, we propose a practical random walk method to simulate microsurface scattering, which accurately solves the derived formulation. Our micrograin BSDF model effectively eliminates the energy loss artifacts inherent in the previous model while significantly reducing noise, providing a physically accurate yet artistically controllable solution for rendering porous materials.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101288"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144771657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-25DOI: 10.1016/j.gmod.2025.101300
Hai Yuan , Xia Yuan , Yanli Liu , Guanyu Xing , Jing Hu , Xi Wu , Zijun Zhou
Virtual human avatars are essential for applications such as gaming, augmented reality, and virtual production. However, existing methods struggle to achieve high fidelity reconstruction from monocular input while keeping hardware costs low. Many approaches rely on the SMPL body prior and apply vertex offsets to represent clothed avatars. Unfortunately, excessive offsets often cause misalignment and blurred contours, particularly around clothing wrinkles, silhouette boundaries, and facial regions. To address these limitations, we propose a dual branch framework for human avatar reconstruction from monocular video. A lightweight Vertex Align Net (VAN) predicts per-vertex normal direction offsets on the SMPL mesh to achieve coarse geometric alignment and guide Gaussian-based human avatar modeling. In parallel, we construct a high resolution facial Gaussian branch based on FLAME estimated parameters, with facial regions localized via pretrained detectors. The facial and body renderings are fused using a semantic mask to enhance facial clarity and ensure globally consistent avatar appearance. Experiments demonstrate that our method surpasses state of the art approaches in modeling animatable human avatars with fine grained fidelity.
{"title":"Adaptive mesh-aligned Gaussian Splatting for monocular human avatar reconstruction","authors":"Hai Yuan , Xia Yuan , Yanli Liu , Guanyu Xing , Jing Hu , Xi Wu , Zijun Zhou","doi":"10.1016/j.gmod.2025.101300","DOIUrl":"10.1016/j.gmod.2025.101300","url":null,"abstract":"<div><div>Virtual human avatars are essential for applications such as gaming, augmented reality, and virtual production. However, existing methods struggle to achieve high fidelity reconstruction from monocular input while keeping hardware costs low. Many approaches rely on the SMPL body prior and apply vertex offsets to represent clothed avatars. Unfortunately, excessive offsets often cause misalignment and blurred contours, particularly around clothing wrinkles, silhouette boundaries, and facial regions. To address these limitations, we propose a dual branch framework for human avatar reconstruction from monocular video. A lightweight Vertex Align Net (VAN) predicts per-vertex normal direction offsets on the SMPL mesh to achieve coarse geometric alignment and guide Gaussian-based human avatar modeling. In parallel, we construct a high resolution facial Gaussian branch based on FLAME estimated parameters, with facial regions localized via pretrained detectors. The facial and body renderings are fused using a semantic mask to enhance facial clarity and ensure globally consistent avatar appearance. Experiments demonstrate that our method surpasses state of the art approaches in modeling animatable human avatars with fine grained fidelity.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101300"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144893147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-27DOI: 10.1016/j.gmod.2025.101299
L. Rocca , F. Iuricich , E. Puppo
We consider 2D scalar fields sampled on a regular grid. When the gradient is low relative to the resolution of the dataset’s range, the signal may contain flat spots: connected areas where all points share the same value. Flat spots hinder certain analyses, such as topological characterization or drainage network computations. We present an algorithm to determine a symbolic slope inside flat spots and consistently place a minimal set of critical points, in a way that is less biased than state-of-the-art methods. We present experimental results on both synthetic and real data, demonstrating how our method provides a more plausible positioning of critical points and a better recovery of the Morse–Smale complex.
{"title":"Disambiguating flat spots in discrete scalar fields","authors":"L. Rocca , F. Iuricich , E. Puppo","doi":"10.1016/j.gmod.2025.101299","DOIUrl":"10.1016/j.gmod.2025.101299","url":null,"abstract":"<div><div>We consider 2D scalar fields sampled on a regular grid. When the gradient is low relative to the resolution of the dataset’s range, the signal may contain <em>flat spots</em>: connected areas where all points share the same value. Flat spots hinder certain analyses, such as topological characterization or drainage network computations. We present an algorithm to determine a symbolic slope inside flat spots and consistently place a minimal set of critical points, in a way that is less biased than state-of-the-art methods. We present experimental results on both synthetic and real data, demonstrating how our method provides a more plausible positioning of critical points and a better recovery of the Morse–Smale complex.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101299"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144904093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-11DOI: 10.1016/j.gmod.2025.101290
Zhifeng Xie , Rui Qiu , Qile He , Mengtian Li , Xin Tan
Scene understanding is a computer vision task that involves grasping the pixel-level distribution of objects. Unlike most research focuses on single-scene models, we consider a more versatile proposal: domain-incremental learning for scene understanding. This allows us to adapt well-studied single-scene models into multi-scene models, reducing data requirements and ensuring model flexibility. However, domain-incremental learning that leverages correlations between scene domains has yet to be explored. To address this challenge, we propose a Domain-Incremental Learning Paradigm (D-ILP) for scene understanding, along with a new strategy of Pseudo-Replay Generation (PRG) that does not require manual labeling. Specifically, D-ILP leverages pre-trained single-scene models and incremental images for supervised training to acquire new knowledge from other scenes. As a pre-trained generation model, PRG can controllably generate pseudo-replays resembling source images from incremental images and text prompts. These pseudo-replays are utilized to minimize catastrophic forgetting in the original scene. We perform experiments with three publicly accessible models: Mask2Former, Segformer, and DeepLabv3+. With successfully transforming these single-scene models into multi-scene models, we achieve high-quality parsing results for original and new scenes simultaneously. Meanwhile, the validity and rationality of our method are proved by the analysis of D-ILP.
{"title":"Domain-Incremental Learning Paradigm for scene understanding via Pseudo-Replay Generation","authors":"Zhifeng Xie , Rui Qiu , Qile He , Mengtian Li , Xin Tan","doi":"10.1016/j.gmod.2025.101290","DOIUrl":"10.1016/j.gmod.2025.101290","url":null,"abstract":"<div><div>Scene understanding is a computer vision task that involves grasping the pixel-level distribution of objects. Unlike most research focuses on single-scene models, we consider a more versatile proposal: domain-incremental learning for scene understanding. This allows us to adapt well-studied single-scene models into multi-scene models, reducing data requirements and ensuring model flexibility. However, domain-incremental learning that leverages correlations between scene domains has yet to be explored. To address this challenge, we propose a Domain-Incremental Learning Paradigm (D-ILP) for scene understanding, along with a new strategy of Pseudo-Replay Generation (PRG) that does not require manual labeling. Specifically, D-ILP leverages pre-trained single-scene models and incremental images for supervised training to acquire new knowledge from other scenes. As a pre-trained generation model, PRG can controllably generate pseudo-replays resembling source images from incremental images and text prompts. These pseudo-replays are utilized to minimize catastrophic forgetting in the original scene. We perform experiments with three publicly accessible models: Mask2Former, Segformer, and DeepLabv3+. With successfully transforming these single-scene models into multi-scene models, we achieve high-quality parsing results for original and new scenes simultaneously. Meanwhile, the validity and rationality of our method are proved by the analysis of D-ILP.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101290"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144810321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-19DOI: 10.1016/j.gmod.2025.101293
Jia-Hong Liu , Shao-Kui Zhang , Shuran Sun , Zihao Wang , Song-Hai Zhang
Recently, indoor scene synthesis has gathered significant attention, leading to the development of numerous indoor datasets. However, existing datasets only address static furniture and scenes, ignoring the need for dynamic interior design scenarios that emphasize flexible functionalities. Addressing this gap, we present DIFF (Dataset for Indoor Flexible Furniture), featuring expertly crafted and labeled furniture modules capable of inter-transforming between different states, e.g., a cabinet can be inter-transformed to a desk. Each module exhibits flexibility in shifting to multiple shapes and functionalities. Additionally, we propose a method that adapts our dataset to generate flexible layouts. By matching our flexible objects to objects from existing datasets, we use a graph-based approach to migrate the spatial relation priors for optimizing a layout; subsequent layouts are then generated by minimizing a transition-cost function. Analyses and user studies validate the quality of our modules and demonstrate the plausibility of the proposed method.
{"title":"DIFF: A dataset for indoor flexible furniture","authors":"Jia-Hong Liu , Shao-Kui Zhang , Shuran Sun , Zihao Wang , Song-Hai Zhang","doi":"10.1016/j.gmod.2025.101293","DOIUrl":"10.1016/j.gmod.2025.101293","url":null,"abstract":"<div><div>Recently, indoor scene synthesis has gathered significant attention, leading to the development of numerous indoor datasets. However, existing datasets only address static furniture and scenes, ignoring the need for dynamic interior design scenarios that emphasize flexible functionalities. Addressing this gap, we present DIFF (Dataset for Indoor Flexible Furniture), featuring expertly crafted and labeled furniture modules capable of inter-transforming between different states, e.g., a cabinet can be inter-transformed to a desk. Each module exhibits flexibility in shifting to multiple shapes and functionalities. Additionally, we propose a method that adapts our dataset to generate flexible layouts. By matching our flexible objects to objects from existing datasets, we use a graph-based approach to migrate the spatial relation priors for optimizing a layout; subsequent layouts are then generated by minimizing a transition-cost function. Analyses and user studies validate the quality of our modules and demonstrate the plausibility of the proposed method.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101293"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144867141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-01Epub Date: 2025-08-21DOI: 10.1016/j.gmod.2025.101284
Hang Zhao , Kexiong Yu , Yuhang Huang , Renjiao Yi , Chenyang Zhu , Kai Xu
Combinatorial Optimization (CO) problems are fundamentally important in numerous real-world applications across diverse industries, notably computer graphics, characterized by entailing enormous solution space and demanding time-sensitive response. Despite recent advancements in neural solvers, their limited expressiveness struggles to capture the multi-modal nature of CO landscapes. While some research has adopted diffusion models, these methods sample solutions indiscriminately from the entire NP-complete solution space with time-consuming denoising processes, limiting scalability for large-scale problems. We propose DISCO, an efficient DIffusion Solver for large-scale Combinatorial Optimization problems that excels in both solution quality and inference speed. DISCO’s efficacy is twofold: First, it enhances solution quality by constraining the sampling space to a more meaningful domain guided by solution residues, while preserving the multi-modal properties of the output distributions. Second, it accelerates the denoising process through an analytically solvable approach, enabling solution sampling with very few reverse-time steps and significantly reducing inference time. This inference-speed advantage is further amplified by Jittor, a high-performance learning framework based on just-in-time compiling and meta-operators. DISCO delivers strong performance on large-scale Traveling Salesman Problems and challenging Maximal Independent Set benchmarks, with inference duration up to 5.38 times faster than existing diffusion solver alternatives. We apply DISCO to design 2D/3D TSP Art, enabling the generation of fluid stroke sequences at reduced path costs. By incorporating DISCO’s multi-modal property into a divide-and-conquer strategy, it can further generalize to solve unseen-scale instances out of the box.
{"title":"DISCO: Efficient Diffusion Solver for large-scale Combinatorial Optimization problems","authors":"Hang Zhao , Kexiong Yu , Yuhang Huang , Renjiao Yi , Chenyang Zhu , Kai Xu","doi":"10.1016/j.gmod.2025.101284","DOIUrl":"10.1016/j.gmod.2025.101284","url":null,"abstract":"<div><div>Combinatorial Optimization (CO) problems are fundamentally important in numerous real-world applications across diverse industries, notably computer graphics, characterized by entailing enormous solution space and demanding time-sensitive response. Despite recent advancements in neural solvers, their limited expressiveness struggles to capture the multi-modal nature of CO landscapes. While some research has adopted diffusion models, these methods sample solutions indiscriminately from the entire NP-complete solution space with time-consuming denoising processes, limiting scalability for large-scale problems. We propose <strong>DISCO</strong>, an efficient <strong>DI</strong>ffusion <strong>S</strong>olver for large-scale <strong>C</strong>ombinatorial <strong>O</strong>ptimization problems that excels in both solution quality and inference speed. DISCO’s efficacy is twofold: First, it enhances solution quality by constraining the sampling space to a more meaningful domain guided by solution residues, while preserving the multi-modal properties of the output distributions. Second, it accelerates the denoising process through an analytically solvable approach, enabling solution sampling with very few reverse-time steps and significantly reducing inference time. This inference-speed advantage is further amplified by Jittor, a high-performance learning framework based on just-in-time compiling and meta-operators. DISCO delivers strong performance on large-scale Traveling Salesman Problems and challenging Maximal Independent Set benchmarks, with inference duration up to 5.38 times faster than existing diffusion solver alternatives. We apply DISCO to design 2D/3D TSP Art, enabling the generation of fluid stroke sequences at reduced path costs. By incorporating DISCO’s multi-modal property into a divide-and-conquer strategy, it can further generalize to solve unseen-scale instances out of the box.</div></div>","PeriodicalId":55083,"journal":{"name":"Graphical Models","volume":"141 ","pages":"Article 101284"},"PeriodicalIF":2.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144878547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}