Despite the remarkable progress in 3D talking head generation, directly generating 3D talking human avatars still suffers from rigid facial expressions, distorted hand textures and out-of-sync lip movements. In this paper, we extend speaker-specific talking head generation task to talking human avatar synthesis and propose a novel pipeline, THGS, that animates lifelike Talking Human avatars using 3D Gaussian Splatting (3DGS). Given speech audio, expression and body poses as input, THGS effectively overcomes the limitations of 3DGS human re-construction methods in capturing expressive dynamics, such as mouth movements, facial expressions and hand gestures, from a short monocular video. Firstly, we introduce a simple yet effective Learnable Expression Blendshapes (LEB) for facial dynamics re-construction, where subtle facial dynamics can be generated by linearly combining the static head model and expression blendshapes. Secondly, a Spatial Audio Attention Module (SAAM) is proposed for lip-synced mouth movement animation, building connections between speech audio and mouth Gaussian movements. Thirdly, we employ a body pose, expression and skinning weights joint optimization strategy to optimize these parameters on the fly, which aligns hand movements and expressions better with video input. Experimental results demonstrate that THGS can achieve high-fidelity 3D talking human avatar animation at 150+ fps on a web-based rendering system, improving the requirements of real-time applications. Our project page is at https://sora158.github.io/THGS.github.io/.
{"title":"THGS: Lifelike Talking Human Avatar Synthesis From Monocular Video Via 3D Gaussian Splatting","authors":"Chuang Chen, Lingyun Yu, Quanwei Yang, Aihua Zheng, Hongtao Xie","doi":"10.1111/cgf.15282","DOIUrl":"https://doi.org/10.1111/cgf.15282","url":null,"abstract":"<p>Despite the remarkable progress in 3D talking head generation, directly generating 3D talking human avatars still suffers from rigid facial expressions, distorted hand textures and out-of-sync lip movements. In this paper, we extend speaker-specific talking head generation task to <b>talking human avatar synthesis</b> and propose a novel pipeline, <i>THGS</i>, that animates lifelike Talking Human avatars using 3D Gaussian Splatting (3DGS). Given speech audio, expression and body poses as input, <i>THGS</i> effectively overcomes the limitations of 3DGS human re-construction methods in capturing expressive dynamics, such as <b>mouth movements, facial expressions and hand gestures</b>, from a short monocular video. Firstly, we introduce a simple yet effective <b>Learnable Expression Blendshapes (LEB)</b> for facial dynamics re-construction, where subtle facial dynamics can be generated by linearly combining the static head model and expression blendshapes. Secondly, a <b>Spatial Audio Attention Module (SAAM)</b> is proposed for lip-synced mouth movement animation, building connections between speech audio and mouth Gaussian movements. Thirdly, we employ a <b>body pose, expression and skinning weights joint optimization strategy</b> to optimize these parameters on the fly, which aligns hand movements and expressions better with video input. Experimental results demonstrate that <i>THGS</i> can achieve high-fidelity 3D talking human avatar animation at 150+ fps on a web-based rendering system, improving the requirements of real-time applications. Our project page is at https://sora158.github.io/THGS.github.io/.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Toshit Jain, Upkar Singh, Varun Singh, Vijay Kumar Boda, Ingrid Hotz, Sathish S. Vadhiyar, P. N. Vinayachandran, Vijay Natarajan
Oceanographers rely on visual analysis to interpret model simulations, identify events and phenomena, and track dynamic ocean processes. The ever increasing resolution and complexity of ocean data due to its dynamic nature and multivariate relationships demands a scalable and adaptable visualization tool for interactive exploration. We introduce pyParaOcean, a scalable and interactive visualization system designed specifically for ocean data analysis. pyParaOcean offers specialized modules for common oceanographic analysis tasks, including eddy identification and salinity movement tracking. These modules seamlessly integrate with ParaView as filters, ensuring a user-friendly and easy-to-use system while leveraging the parallelization capabilities of ParaView and a plethora of inbuilt general-purpose visualization functionalities. The creation of an auxiliary dataset stored as a Cinema database helps address I/O and network bandwidth bottlenecks while supporting the generation of quick overview visualizations. We present a case study on the Bay of Bengal to demonstrate the utility of the system and scaling studies to evaluate the efficiency of the system.
{"title":"A Scalable System for Visual Analysis of Ocean Data","authors":"Toshit Jain, Upkar Singh, Varun Singh, Vijay Kumar Boda, Ingrid Hotz, Sathish S. Vadhiyar, P. N. Vinayachandran, Vijay Natarajan","doi":"10.1111/cgf.15279","DOIUrl":"https://doi.org/10.1111/cgf.15279","url":null,"abstract":"<p>Oceanographers rely on visual analysis to interpret model simulations, identify events and phenomena, and track dynamic ocean processes. The ever increasing resolution and complexity of ocean data due to its dynamic nature and multivariate relationships demands a scalable and adaptable visualization tool for interactive exploration. We introduce pyParaOcean, a scalable and interactive visualization system designed specifically for ocean data analysis. pyParaOcean offers specialized modules for common oceanographic analysis tasks, including eddy identification and salinity movement tracking. These modules seamlessly integrate with ParaView as filters, ensuring a user-friendly and easy-to-use system while leveraging the parallelization capabilities of ParaView and a plethora of inbuilt general-purpose visualization functionalities. The creation of an auxiliary dataset stored as a Cinema database helps address I/O and network bandwidth bottlenecks while supporting the generation of quick overview visualizations. We present a case study on the Bay of Bengal to demonstrate the utility of the system and scaling studies to evaluate the efficiency of the system.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jan Dvořák, Filip Hácha, Gerasimos Arvanitis, David Podgorelec, Konstantinos Moustakas, Libor Váša
Time-varying meshes (TVMs), that is mesh sequences with varying connectivity, are a greatly versatile representation of shapes evolving in time, as they allow a surface topology to change or details to appear or disappear at any time during the sequence. This, however, comes at the cost of large storage size. Since 2003, there have been attempts to compress such data efficiently. While the problem may seem trivial at first sight, considering the strong temporal coherence of shapes represented by the individual frames, it turns out that the varying connectivity and the absence of implicit correspondence information that stems from it makes it rather difficult to exploit the redundancies present in the data. Therefore, efficient and general TVM compression is still considered an open problem. We describe and categorize existing approaches while pointing out the current challenges in the field and hint at some related techniques that might be helpful in addressing them. We also provide an overview of the reported performance of the discussed methods and a list of datasets that are publicly available for experiments. Finally, we also discuss potential future trends in the field.
{"title":"Survey of Inter-Prediction Methods for Time-Varying Mesh Compression","authors":"Jan Dvořák, Filip Hácha, Gerasimos Arvanitis, David Podgorelec, Konstantinos Moustakas, Libor Váša","doi":"10.1111/cgf.15278","DOIUrl":"https://doi.org/10.1111/cgf.15278","url":null,"abstract":"<p>Time-varying meshes (TVMs), that is mesh sequences with varying connectivity, are a greatly versatile representation of shapes evolving in time, as they allow a surface topology to change or details to appear or disappear at any time during the sequence. This, however, comes at the cost of large storage size. Since 2003, there have been attempts to compress such data efficiently. While the problem may seem trivial at first sight, considering the strong temporal coherence of shapes represented by the individual frames, it turns out that the varying connectivity and the absence of implicit correspondence information that stems from it makes it rather difficult to exploit the redundancies present in the data. Therefore, efficient and general TVM compression is still considered an open problem. We describe and categorize existing approaches while pointing out the current challenges in the field and hint at some related techniques that might be helpful in addressing them. We also provide an overview of the reported performance of the discussed methods and a list of datasets that are publicly available for experiments. Finally, we also discuss potential future trends in the field.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15278","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Khanteimouri P., Mandad M., Campen M.: Rational Bézier Guarding. Computer Graphics Forum 41, 5 (2022), 89–99.
The proof of Lemma 5 contains a mistake; while various kinds of variation diminishing properties of Bézier triangles are known, one of the specific geometric kind assumed is not available. The lemma does not hold in full generality as stated, for arbitrary Bézier triangles. Whether it holds for the restricted class of Bézier triangles constructed by the method is an open question, no counterexample is known.
We apologize for this error.
{"title":"Erratum to “Rational Bézier Guarding”","authors":"","doi":"10.1111/cgf.15277","DOIUrl":"https://doi.org/10.1111/cgf.15277","url":null,"abstract":"<p>Khanteimouri P., Mandad M., Campen M.: Rational Bézier Guarding. <i>Computer Graphics Forum 41</i>, 5 (2022), 89–99.</p><p>The proof of Lemma 5 contains a mistake; while various kinds of variation diminishing properties of Bézier triangles are known, one of the specific geometric kind assumed is not available. The lemma does not hold in full generality as stated, for arbitrary Bézier triangles. Whether it holds for the restricted class of Bézier triangles constructed by the method is an open question, no counterexample is known.</p><p>We apologize for this error.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15277","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The hybrid Lagrangian/Eulerian formulation of continuum shells is highly effective for producing challenging simulations of thin materials like cloth with bending resistance and frictional contact. However, existing formulations are restricted to materials that do not undergo tearing nor fracture due to the difficulties associated with incorporating strong discontinuities of field quantities like velocity via basis enrichment while maintaining continuity or regularity. We propose an extension of this formulation to simulate dynamic tearing and fracturing of thin shells using Kirchhoff–Love continuum theory. Damage, which manifests as cracks or tears, is propagated by tracking the evolution of a time-dependent phase-field in the co-dimensional manifold, where a moving least-squares (MLS) approximation then captures the strong discontinuities of interpolated field quantities near the crack. Our approach is capable of simulating challenging scenarios of this tearing and fracture, all-the-while harnessing the existing benefits of the hybrid Lagrangian/Eulerian formulation to expand the domain of possible effects. The method is also amenable to user-guided control, which serves to influence the propagation of cracks or tears such that they follow prescribed paths during simulation.
{"title":"A Hybrid Lagrangian–Eulerian Formulation of Thin-Shell Fracture","authors":"L. Fan, F. M. Chitalu, T. Komura","doi":"10.1111/cgf.15273","DOIUrl":"https://doi.org/10.1111/cgf.15273","url":null,"abstract":"<p>The hybrid Lagrangian/Eulerian formulation of continuum shells is highly effective for producing challenging simulations of thin materials like cloth with bending resistance and frictional contact. However, existing formulations are restricted to materials that do not undergo tearing nor fracture due to the difficulties associated with incorporating strong discontinuities of field quantities like velocity via basis enrichment while maintaining <span></span><math></math> continuity or <span></span><math></math> regularity. We propose an extension of this formulation to simulate dynamic tearing and fracturing of thin shells using Kirchhoff–Love continuum theory. Damage, which manifests as cracks or tears, is propagated by tracking the evolution of a time-dependent phase-field in the co-dimensional manifold, where a moving least-squares (MLS) approximation then captures the strong discontinuities of interpolated field quantities near the crack. Our approach is capable of simulating challenging scenarios of this tearing and fracture, all-the-while harnessing the existing benefits of the hybrid Lagrangian/Eulerian formulation to expand the domain of possible effects. The method is also amenable to user-guided control, which serves to influence the propagation of cracks or tears such that they follow prescribed paths during simulation.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15273","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Hartwig, C. v. Onzenoodt, D. Engel, P. Hermosilla, T. Ropinski
Cluster separation is a task typically tackled by widely used clustering techniques, such as k-means or DBSCAN. However, these algorithms are based on non-perceptual metrics, and our experiments demonstrate that their output does not reflect human cluster perception. To bridge the gap between human cluster perception and machine-computed clusters, we propose HPSCAN, a learning strategy that operates directly on scattered data. To learn perceptual cluster separation on such data, we crowdsourced the labeling of bivariate (scatterplot) datasets to 384 human participants. We train our HPSCAN model on these human-annotated data. Instead of rendering these data as scatterplot images, we used their x and y point coordinates as input to a modified PointNet++ architecture, enabling direct inference on point clouds. In this work, we provide details on how we collected our dataset, report statistics of the resulting annotations, and investigate the perceptual agreement of cluster separation for real-world data. We also report the training and evaluation protocol for HPSCAN and introduce a novel metric, that measures the accuracy between a clustering technique and a group of human annotators. We explore predicting point-wise human agreement to detect ambiguities. Finally, we compare our approach to 10 established clustering techniques and demonstrate that HPSCAN is capable of generalizing to unseen and out-of-scope data.
{"title":"HPSCAN: Human Perception-Based Scattered Data Clustering","authors":"S. Hartwig, C. v. Onzenoodt, D. Engel, P. Hermosilla, T. Ropinski","doi":"10.1111/cgf.15275","DOIUrl":"https://doi.org/10.1111/cgf.15275","url":null,"abstract":"<p>Cluster separation is a task typically tackled by widely used clustering techniques, such as k-means or DBSCAN. However, these algorithms are based on non-perceptual metrics, and our experiments demonstrate that their output does not reflect human cluster perception. To bridge the gap between human cluster perception and machine-computed clusters, we propose HPSCAN, a learning strategy that operates directly on scattered data. To learn perceptual cluster separation on such data, we crowdsourced the labeling of <span></span><math></math> bivariate (scatterplot) datasets to 384 human participants. We train our HPSCAN model on these human-annotated data. Instead of rendering these data as scatterplot images, we used their <i>x</i> and <i>y</i> point coordinates as input to a modified PointNet++ architecture, enabling direct inference on point clouds. In this work, we provide details on how we collected our dataset, report statistics of the resulting annotations, and investigate the perceptual agreement of cluster separation for real-world data. We also report the training and evaluation protocol for HPSCAN and introduce a novel metric, that measures the accuracy between a clustering technique and a group of human annotators. We explore predicting point-wise human agreement to detect ambiguities. Finally, we compare our approach to 10 established clustering techniques and demonstrate that HPSCAN is capable of generalizing to unseen and out-of-scope data.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15275","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Novel view synthesis for talking heads presents significant challenges due to the complex and diverse motion transformations involved. Conventional methods often resort to reliance on structure priors, like facial templates, to warp observed images into a canonical space conducive to rendering. However, the incorporation of such priors introduces a trade-off-while aiding in synthesis, they concurrently amplify model complexity, limiting generalizability to other deformable scenes. Departing from this paradigm, we introduce a pioneering solution: the motion-conditioned neural radiance field, MoNeRF, designed to model talking heads through latent motion navigation. At the core of MoNeRF lies a novel approach utilizing a compact set of latent codes to represent orthogonal motion directions. This innovative strategy empowers MoNeRF to efficiently capture and depict intricate scene motion by linearly combining these latent codes. In an extended capability, MoNeRF facilitates motion control through latent code adjustments, supports view transfer based on reference videos, and seamlessly extends its applicability to model human bodies without necessitating structural modifications. Rigorous quantitative and qualitative experiments unequivocally demonstrate MoNeRF's superior performance compared to state-of-the-art methods in talking head synthesis. We will release the source code upon publication.
{"title":"MoNeRF: Deformable Neural Rendering for Talking Heads via Latent Motion Navigation","authors":"X. Li, Y. Ding, R. Li, Z. Tang, K. Li","doi":"10.1111/cgf.15274","DOIUrl":"https://doi.org/10.1111/cgf.15274","url":null,"abstract":"<p>Novel view synthesis for talking heads presents significant challenges due to the complex and diverse motion transformations involved. Conventional methods often resort to reliance on structure priors, like facial templates, to warp observed images into a canonical space conducive to rendering. However, the incorporation of such priors introduces a trade-off-while aiding in synthesis, they concurrently amplify model complexity, limiting generalizability to other deformable scenes. Departing from this paradigm, we introduce a pioneering solution: the motion-conditioned neural radiance field, MoNeRF, designed to model talking heads through latent motion navigation. At the core of MoNeRF lies a novel approach utilizing a compact set of latent codes to represent orthogonal motion directions. This innovative strategy empowers MoNeRF to efficiently capture and depict intricate scene motion by linearly combining these latent codes. In an extended capability, MoNeRF facilitates motion control through latent code adjustments, supports view transfer based on reference videos, and seamlessly extends its applicability to model human bodies without necessitating structural modifications. Rigorous quantitative and qualitative experiments unequivocally demonstrate MoNeRF's superior performance compared to state-of-the-art methods in talking head synthesis. We will release the source code upon publication.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513626","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fernando Yanez, Cristina Conati, Alvitta Ottley, Carolina Nobre
Research shows that user traits can modulate the use of visualization systems and have a measurable influence on users' accuracy, speed, and attention when performing visual analysis. This highlights the importance of user-adaptive visualization that can modify themselves to the characteristics and preferences of the user. However, there are very few such visualization systems, as creating them requires broad knowledge from various sub-domains of the visualization community. A user-adaptive system must consider which user traits they adapt to, their adaptation logic and the types of interventions they support. In this STAR, we survey a broad space of existing literature and consolidate them to structure the process of creating user-adaptive visualizations into five components: Capture Ⓐ Input from the user and any relevant peripheral information. Perform computational Ⓑ User Modelling with this input to construct a Ⓒ User Representation. Employ Ⓓ Adaptation Assignment logic to identify when and how to introduce Ⓔ Interventions. Our novel taxonomy provides a road map for work in this area, describing the rich space of current approaches and highlighting open areas for future work.
{"title":"The State of the Art in User-Adaptive Visualizations","authors":"Fernando Yanez, Cristina Conati, Alvitta Ottley, Carolina Nobre","doi":"10.1111/cgf.15271","DOIUrl":"https://doi.org/10.1111/cgf.15271","url":null,"abstract":"<p>Research shows that user traits can modulate the use of visualization systems and have a measurable influence on users' accuracy, speed, and attention when performing visual analysis. This highlights the importance of user-adaptive visualization that can modify themselves to the characteristics and preferences of the user. However, there are very few such visualization systems, as creating them requires broad knowledge from various sub-domains of the visualization community. A user-adaptive system must consider which user traits they adapt to, their adaptation logic and the types of interventions they support. In this STAR, we survey a broad space of existing literature and consolidate them to structure the process of creating user-adaptive visualizations into five components: Capture Ⓐ <b>Input</b> from the user and any relevant peripheral information. Perform computational Ⓑ <b>User Modelling</b> with this input to construct a Ⓒ <b>User Representation</b>. Employ Ⓓ <b>Adaptation Assignment</b> logic to identify when and how to introduce Ⓔ <b>Interventions</b>. Our novel taxonomy provides a road map for work in this area, describing the rich space of current approaches and highlighting open areas for future work.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15271","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
User confidence plays an important role in guided visual data analysis scenarios, especially when uncertainty is involved in the analytical process. However, measuring confidence in practical scenarios remains an open challenge, as previous work relies primarily on self-reporting methods. In this work, we propose a quantitative approach to measure user confidence—as opposed to trust—in an analytical scenario. We do so by exploiting the respective user interaction provenance graph and examining the impact of guidance using a set of network metrics. We assess the usefulness of our proposed metrics through a user study that correlates results obtained from self-reported confidence assessments and our metrics—both with and without guidance. The results suggest that our metrics improve the evaluation of user confidence compared to available approaches. In particular, we found a correlation between self-reported confidence and some of the proposed provenance network metrics. The quantitative results, though, do not show a statistically significant impact of the guidance on user confidence. An additional descriptive analysis suggests that guidance could impact users' confidence and that the qualitative analysis of the provenance network topology can provide a comprehensive view of changes in user confidence. Our results indicate that our proposed metrics and the provenance network graph representation support the evaluation of user confidence and, subsequently, the effective development of guidance in VA.
{"title":"ConAn: Measuring and Evaluating User Confidence in Visual Data Analysis Under Uncertainty","authors":"M. Musleh, D. Ceneda, H. Ehlers, R. G. Raidou","doi":"10.1111/cgf.15272","DOIUrl":"https://doi.org/10.1111/cgf.15272","url":null,"abstract":"<p>User confidence plays an important role in guided visual data analysis scenarios, especially when uncertainty is involved in the analytical process. However, measuring confidence in practical scenarios remains an open challenge, as previous work relies primarily on self-reporting methods. In this work, we propose a quantitative approach to measure user confidence—as opposed to trust—in an analytical scenario. We do so by exploiting the respective user interaction provenance graph and examining the impact of guidance using a set of network metrics. We assess the usefulness of our proposed metrics through a user study that correlates results obtained from self-reported confidence assessments and our metrics—both with and without guidance. The results suggest that our metrics improve the evaluation of user confidence compared to available approaches. In particular, we found a correlation between self-reported confidence and some of the proposed provenance network metrics. The quantitative results, though, do not show a statistically significant impact of the guidance on user confidence. An additional descriptive analysis suggests that guidance could impact users' confidence and that the qualitative analysis of the provenance network topology can provide a comprehensive view of changes in user confidence. Our results indicate that our proposed metrics and the provenance network graph representation support the evaluation of user confidence and, subsequently, the effective development of guidance in VA.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.15272","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143513879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}