Pub Date : 2024-03-01DOI: 10.1016/j.visinf.2023.06.008
Ying Zhao , Shenglan Lv , Wenwei Long , Yilun Fan , Jian Yuan , Haojin Jiang , Fangfang Zhou
Malicious webshells currently present tremendous threats to cloud security. Most relevant studies and open webshell datasets consider malicious webshell defense as a binary classification problem, that is, identifying whether a webshell is malicious or benign. However, a fine-grained multi-classification is urgently needed to enable precise responses and active defenses on malicious webshell threats. This paper introduces a malicious webshell family dataset named MWF to facilitate webshell multi-classification researches. This dataset contains 1359 malicious webshell samples originally obtained from the cloud servers of Alibaba Cloud. Each of them is provided with a family label. The samples of the same family generally present similar characteristics or behaviors. The dataset has a total of 78 families and 22 outliers. Moreover, this paper introduces the human–machine collaboration process that is adopted to remove benign or duplicate samples, address privacy issues, and determine the family of each sample. This paper also compares the distinguished features of the MWF dataset with previous datasets and summarizes the potential applied areas in cloud security and generalized sequence, graph, and tree data analytics and visualization.
{"title":"Malicious webshell family dataset for webshell multi-classification research","authors":"Ying Zhao , Shenglan Lv , Wenwei Long , Yilun Fan , Jian Yuan , Haojin Jiang , Fangfang Zhou","doi":"10.1016/j.visinf.2023.06.008","DOIUrl":"10.1016/j.visinf.2023.06.008","url":null,"abstract":"<div><p>Malicious webshells currently present tremendous threats to cloud security. Most relevant studies and open webshell datasets consider malicious webshell defense as a binary classification problem, that is, identifying whether a webshell is malicious or benign. However, a fine-grained multi-classification is urgently needed to enable precise responses and active defenses on malicious webshell threats. This paper introduces a malicious webshell family dataset named MWF to facilitate webshell multi-classification researches. This dataset contains 1359 malicious webshell samples originally obtained from the cloud servers of Alibaba Cloud. Each of them is provided with a family label. The samples of the same family generally present similar characteristics or behaviors. The dataset has a total of 78 families and 22 outliers. Moreover, this paper introduces the human–machine collaboration process that is adopted to remove benign or duplicate samples, address privacy issues, and determine the family of each sample. This paper also compares the distinguished features of the MWF dataset with previous datasets and summarizes the potential applied areas in cloud security and generalized sequence, graph, and tree data analytics and visualization.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000335/pdfft?md5=0e04b6b31402572c03a419f9b7597a47&pid=1-s2.0-S2468502X23000335-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75769287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.visinf.2023.11.001
Jun Wang , Bohan Lei , Liya Ding , Xiaoyin Xu , Xianfeng Gu , Min Zhang
Medical image generation has recently garnered significant interest among researchers. However, the primary generative models, such as Generative Adversarial Networks (GANs), often encounter challenges during training, including mode collapse. To address these issues, we proposed the AE-COT-GAN model (Autoencoder-based Conditional Optimal Transport Generative Adversarial Network) for the generation of medical images belonging to specific categories. The training process of our model comprises three fundamental components. The training process of our model encompasses three fundamental components. First, we employ an autoencoder model to obtain a low-dimensional manifold representation of real images. Second, we apply extended semi-discrete optimal transport to map Gaussian noise distribution to the latent space distribution and obtain corresponding labels effectively. This procedure leads to the generation of new latent codes with known labels. Finally, we integrate a GAN to train the decoder further to generate medical images. To evaluate the performance of the AE-COT-GAN model, we conducted experiments on two medical image datasets, namely DermaMNIST and BloodMNIST. The model’s performance was compared with state-of-the-art generative models. Results show that the AE-COT-GAN model had excellent performance in generating medical images. Moreover, it effectively addressed the common issues associated with traditional GANs.
医学图像生成最近引起了研究人员的极大兴趣。然而,主要的生成模型,如生成对抗网络(GAN),在训练过程中经常会遇到模式崩溃等挑战。为了解决这些问题,我们提出了 AE-COT-GAN 模型(基于自动编码器的条件优化传输生成对抗网络),用于生成属于特定类别的医学图像。我们模型的训练过程包括三个基本组成部分。我们模型的训练过程包括三个基本组成部分。首先,我们采用自动编码器模型获得真实图像的低维流形表示。其次,我们应用扩展的半离散最优传输将高斯噪声分布映射到潜空间分布,并有效地获得相应的标签。这一过程可生成带有已知标签的新潜码。最后,我们整合了一个 GAN 来进一步训练解码器,以生成医学图像。为了评估 AE-COT-GAN 模型的性能,我们在两个医学图像数据集(即 DermaMNIST 和 BloodMNIST)上进行了实验。我们将该模型的性能与最先进的生成模型进行了比较。结果表明,AE-COT-GAN 模型在生成医学图像方面表现出色。此外,它还有效地解决了与传统 GAN 相关的常见问题。
{"title":"Autoencoder-based conditional optimal transport generative adversarial network for medical image generation","authors":"Jun Wang , Bohan Lei , Liya Ding , Xiaoyin Xu , Xianfeng Gu , Min Zhang","doi":"10.1016/j.visinf.2023.11.001","DOIUrl":"https://doi.org/10.1016/j.visinf.2023.11.001","url":null,"abstract":"<div><p>Medical image generation has recently garnered significant interest among researchers. However, the primary generative models, such as Generative Adversarial Networks (GANs), often encounter challenges during training, including mode collapse. To address these issues, we proposed the AE-COT-GAN model (Autoencoder-based Conditional Optimal Transport Generative Adversarial Network) for the generation of medical images belonging to specific categories. The training process of our model comprises three fundamental components. The training process of our model encompasses three fundamental components. First, we employ an autoencoder model to obtain a low-dimensional manifold representation of real images. Second, we apply extended semi-discrete optimal transport to map Gaussian noise distribution to the latent space distribution and obtain corresponding labels effectively. This procedure leads to the generation of new latent codes with known labels. Finally, we integrate a GAN to train the decoder further to generate medical images. To evaluate the performance of the AE-COT-GAN model, we conducted experiments on two medical image datasets, namely DermaMNIST and BloodMNIST. The model’s performance was compared with state-of-the-art generative models. Results show that the AE-COT-GAN model had excellent performance in generating medical images. Moreover, it effectively addressed the common issues associated with traditional GANs.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000529/pdfft?md5=3af566b28e15f895521e10dc5d8d1dbc&pid=1-s2.0-S2468502X23000529-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140339084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.visinf.2023.11.003
Jialu Dong , Huijie Zhang , Meiqi Cui , Yiming Lin , Hsiang-Yun Wu , Chongke Bi
Traffic congestion is becoming increasingly severe as a result of urbanization, which not only impedes people’s ability to travel but also hinders the economic development of cities. Modeling the correlation between congestion and its influencing factors using machine learning methods makes it possible to quickly identify congested road segments. Due to the intrinsic black-box character of machine learning models, it is difficult for experts to trust the decision results of road congestion prediction models and understand the significance of congestion-causing factors. In this paper, we present a model interpretability method to investigate the potential causes of traffic congestion and quantify the importance of various influencing factors using the SHAP method. Due to the multidimensionality of these factors, it can be challenging to visually represent the impact of all factors. In response, we propose TCEVis, an interactive visual analytics system that enables multi-level exploration of road conditions. Through three case studies utilizing actual data, we demonstrate that the TCEVis system offers advantages for assisting traffic managers in analyzing the causes of traffic congestion and elucidating the significance of various influencing factors.
{"title":"TCEVis: Visual analytics of traffic congestion influencing factors based on explainable machine learning","authors":"Jialu Dong , Huijie Zhang , Meiqi Cui , Yiming Lin , Hsiang-Yun Wu , Chongke Bi","doi":"10.1016/j.visinf.2023.11.003","DOIUrl":"https://doi.org/10.1016/j.visinf.2023.11.003","url":null,"abstract":"<div><p>Traffic congestion is becoming increasingly severe as a result of urbanization, which not only impedes people’s ability to travel but also hinders the economic development of cities. Modeling the correlation between congestion and its influencing factors using machine learning methods makes it possible to quickly identify congested road segments. Due to the intrinsic black-box character of machine learning models, it is difficult for experts to trust the decision results of road congestion prediction models and understand the significance of congestion-causing factors. In this paper, we present a model interpretability method to investigate the potential causes of traffic congestion and quantify the importance of various influencing factors using the SHAP method. Due to the multidimensionality of these factors, it can be challenging to visually represent the impact of all factors. In response, we propose TCEVis, an interactive visual analytics system that enables multi-level exploration of road conditions. Through three case studies utilizing actual data, we demonstrate that the TCEVis system offers advantages for assisting traffic managers in analyzing the causes of traffic congestion and elucidating the significance of various influencing factors.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000530/pdfft?md5=71c05bc362850cbe9f83fb75c6e85e7f&pid=1-s2.0-S2468502X23000530-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140341083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.visinf.2024.01.002
Praveen Soni , Cyril de Runz , Fatma Bouali , Gilles Venturini
This paper presents a survey on automatic or semi-automatic recommendation systems that help users create dashboards. It starts by showing the important role that dashboards play in data science, and give an informal definition of dashboards, i.e., a set of visualizations possibly with linkage, a screen layout and user feedback. We are mainly interested in systems that use a fully or partially automatic mechanism to recommend dashboards to users. This automation includes the suggestion of data and visualizations, the optimization of the layout and the use of user feedback. We position our work with respect to existing surveys. Starting from a set of over 1000 papers, we have selected and analyzed 19 papers/systems along several dimensions. The main dimensions were the set of considered visualizations, the suggestion method, the utility/objective functions, the layout, and the user interface. We conclude by highlighting the main achievements in this domain and by proposing perspectives.
{"title":"A survey on automatic dashboard recommendation systems","authors":"Praveen Soni , Cyril de Runz , Fatma Bouali , Gilles Venturini","doi":"10.1016/j.visinf.2024.01.002","DOIUrl":"10.1016/j.visinf.2024.01.002","url":null,"abstract":"<div><p>This paper presents a survey on automatic or semi-automatic recommendation systems that help users create dashboards. It starts by showing the important role that dashboards play in data science, and give an informal definition of dashboards, i.e., a set of visualizations possibly with linkage, a screen layout and user feedback. We are mainly interested in systems that use a fully or partially automatic mechanism to recommend dashboards to users. This automation includes the suggestion of data and visualizations, the optimization of the layout and the use of user feedback. We position our work with respect to existing surveys. Starting from a set of over 1000 papers, we have selected and analyzed 19 papers/systems along several dimensions. The main dimensions were the set of considered visualizations, the suggestion method, the utility/objective functions, the layout, and the user interface. We conclude by highlighting the main achievements in this domain and by proposing perspectives.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000032/pdfft?md5=4f07a8572973789f180c169f5da61d6c&pid=1-s2.0-S2468502X24000032-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139631923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.visinf.2024.01.001
Yunpeng Chen , Ying Zhao , Xuanjing Li , Jiang Zhang , Jiang Long , Fangfang Zhou
Data have become valuable assets for enterprises. Data governance aims to manage and reuse data assets, facilitating enterprise management and enabling product innovations. A data lineage graph (DLG) is an abstracted collection of data assets and their data lineages in data governance. Analyzing DLGs can provide rich data insights for data governance. However, the progress of data governance technologies is hindered by the shortage of available open datasets for DLGs. This paper introduces an open dataset of DLGs, including the DLG model, the dataset construction process, and applied areas. This real-world dataset is sourced from Huawei Cloud Computing Technology Company Limited, which contains 18 DLGs with three types of data assets and two types of relations. To the best of our knowledge, this dataset is the first open dataset of DLGs for data governance. This dataset can also support the development of other application areas, such as graph analytics and visualization.
{"title":"An open dataset of data lineage graphs for data governance research","authors":"Yunpeng Chen , Ying Zhao , Xuanjing Li , Jiang Zhang , Jiang Long , Fangfang Zhou","doi":"10.1016/j.visinf.2024.01.001","DOIUrl":"10.1016/j.visinf.2024.01.001","url":null,"abstract":"<div><p>Data have become valuable assets for enterprises. Data governance aims to manage and reuse data assets, facilitating enterprise management and enabling product innovations. A data lineage graph (DLG) is an abstracted collection of data assets and their data lineages in data governance. Analyzing DLGs can provide rich data insights for data governance. However, the progress of data governance technologies is hindered by the shortage of available open datasets for DLGs. This paper introduces an open dataset of DLGs, including the DLG model, the dataset construction process, and applied areas. This real-world dataset is sourced from Huawei Cloud Computing Technology Company Limited, which contains 18 DLGs with three types of data assets and two types of relations. To the best of our knowledge, this dataset is the first open dataset of DLGs for data governance. This dataset can also support the development of other application areas, such as graph analytics and visualization.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X24000020/pdfft?md5=e1a773bd989e7966f5523d7492f34ffd&pid=1-s2.0-S2468502X24000020-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139636525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01DOI: 10.1016/j.visinf.2023.10.005
Jianheng Xiang
As computer graphics technology supports pursuing a photorealistic style, replicated artworks with a photorealistic style overwhelmingly predominate in the computer-generated art circle. Along with the progression of generative technology, this trend may make generative art a virtual world of photorealistic fake, in which the single criterion of expressive style imperils art into the context of a single boring stereotype. This article focuses on the issue of style diversity and its technical feasibility by artistic experiments of generating flower images in StyleGAN. The author insisted that photo both technology and artistic style should not be confined merely for realistic purposes. This proposition was validated in the GAN generation experiment by changing the training materials.
{"title":"On generated artistic styles: Image generation experiments with GAN algorithms","authors":"Jianheng Xiang","doi":"10.1016/j.visinf.2023.10.005","DOIUrl":"10.1016/j.visinf.2023.10.005","url":null,"abstract":"<div><p>As computer graphics technology supports pursuing a photorealistic style, replicated artworks with a photorealistic style overwhelmingly predominate in the computer-generated art circle. Along with the progression of generative technology, this trend may make generative art a virtual world of photorealistic fake, in which the single criterion of expressive style imperils art into the context of a single boring stereotype. This article focuses on the issue of style diversity and its technical feasibility by artistic experiments of generating flower images in StyleGAN. The author insisted that photo both technology and artistic style should not be confined merely for realistic purposes. This proposition was validated in the GAN generation experiment by changing the training materials.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000505/pdfft?md5=0d5b9b7ecd516c8f2a86795017a23204&pid=1-s2.0-S2468502X23000505-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135614582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01DOI: 10.1016/j.visinf.2023.10.001
Tiemeng Li , Songqian Wu , Yanning Jin , Haopai Shi , Shiran Liu
Mixed reality offers a larger visualization space and more intuitive means of interaction for data exploration, and many works have been dedicated to combining 2D visualizations on screen with mixe reality. However, for each combination, we need to customize the implementation of the corresponding mixed reality 3D visualization. It is a challenge to simplify this development process and enable agile building of mixed reality 3D visualizations for 2D visualizations. In addition, many existing 2D visualizations do not provide interfaces oriented to immersive analytics, so how to extend the mixed reality 3D space from existing 2D visualizations is another challenge. This work presents an agile and flexible approach to interactively transfer visualizations from 2D screens to mixed reality 3D spaces. We designed an interactive process for spatial generation of mixed-reality 3D visualizations, defined a unified data transfer framework, integrated data deconstruction techniques for 2D visualizations, implemented interfaces to immersive visualization building tool-kits, and encapsulated these techniques into a tool named X-Space. We validated that the approach is feasible and effective through 2D visualization cases including scatter plots, stacked bar charts, and adjacency matrix. Finally, we conducted expert interviews to discuss the usability and value of the method.
{"title":"X-Space: Interaction design of extending mixed reality space from Web2D visualization","authors":"Tiemeng Li , Songqian Wu , Yanning Jin , Haopai Shi , Shiran Liu","doi":"10.1016/j.visinf.2023.10.001","DOIUrl":"10.1016/j.visinf.2023.10.001","url":null,"abstract":"<div><p>Mixed reality offers a larger visualization space and more intuitive means of interaction for data exploration, and many works have been dedicated to combining 2D visualizations on screen with mixe reality. However, for each combination, we need to customize the implementation of the corresponding mixed reality 3D visualization. It is a challenge to simplify this development process and enable agile building of mixed reality 3D visualizations for 2D visualizations. In addition, many existing 2D visualizations do not provide interfaces oriented to immersive analytics, so how to extend the mixed reality 3D space from existing 2D visualizations is another challenge. This work presents an agile and flexible approach to interactively transfer visualizations from 2D screens to mixed reality 3D spaces. We designed an interactive process for spatial generation of mixed-reality 3D visualizations, defined a unified data transfer framework, integrated data deconstruction techniques for 2D visualizations, implemented interfaces to immersive visualization building tool-kits, and encapsulated these techniques into a tool named X-Space. We validated that the approach is feasible and effective through 2D visualization cases including scatter plots, stacked bar charts, and adjacency matrix. Finally, we conducted expert interviews to discuss the usability and value of the method.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000463/pdfft?md5=327ef25e16772308cc175a2c6dfa9aec&pid=1-s2.0-S2468502X23000463-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135654708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01DOI: 10.1016/j.visinf.2023.07.003
Rusheng Pan , Yunhai Wang , Jiashun Sun , Hongbo Liu , Ying Zhao , Jiazhi Xia , Wei Chen
One main challenge for simplifying node-link diagrams of large-scale social networks lies in that simplified graphs generally contain dense subgroups or cohesive subgraphs. Graph triangles quantify the solid and stable relationships that maintain cohesive subgraphs. Understanding the mechanism of triangles within cohesive subgraphs contributes to illuminating patterns of connections within social networks. However, prior works can hardly handle and visualize triangles in cohesive subgraphs. In this paper, we propose a triangle-based graph simplification approach that can filter and visualize cohesive subgraphs by leveraging a triangle-connectivity called -truss and a force-directed algorithm. We design and implement TriGraph, a web-based visual interface that provides detailed information for exploring and analyzing social networks. Quantitative comparisons with existing methods, two case studies on real-world datasets, and feedback from domain experts demonstrate the effectiveness of TriGraph.
{"title":"Simplifying social networks via triangle-based cohesive subgraphs","authors":"Rusheng Pan , Yunhai Wang , Jiashun Sun , Hongbo Liu , Ying Zhao , Jiazhi Xia , Wei Chen","doi":"10.1016/j.visinf.2023.07.003","DOIUrl":"10.1016/j.visinf.2023.07.003","url":null,"abstract":"<div><p>One main challenge for simplifying node-link diagrams of large-scale social networks lies in that simplified graphs generally contain dense subgroups or cohesive subgraphs. Graph triangles quantify the solid and stable relationships that maintain cohesive subgraphs. Understanding the mechanism of triangles within cohesive subgraphs contributes to illuminating patterns of connections within social networks. However, prior works can hardly handle and visualize triangles in cohesive subgraphs. In this paper, we propose a triangle-based graph simplification approach that can filter and visualize cohesive subgraphs by leveraging a triangle-connectivity called <span><math><mi>k</mi></math></span>-truss and a force-directed algorithm. We design and implement TriGraph, a web-based visual interface that provides detailed information for exploring and analyzing social networks. Quantitative comparisons with existing methods, two case studies on real-world datasets, and feedback from domain experts demonstrate the effectiveness of TriGraph.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000347/pdfft?md5=9f8781159cb5084af8adc00c529e2a37&pid=1-s2.0-S2468502X23000347-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72536742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01DOI: 10.1016/j.visinf.2023.10.004
Yi Wu , Minghong Zheng , Changqing Weng
Art therapy as an intervention has been shown to alleviate social impairment in people with AD. Meanwhile, digital technology (DTS) has been shown to perform well in different degenerative dementias through mobile devices and apps. However, it is unclear whether digital art creation therapy has an impact on the speech function of people with early AD. Therefore, the aim of this study was to confirm whether digital art creation therapy has an ameliorating effect on language decline in AD patients through the KnowU social teleprompter. This study was a controlled trial in which 16 patients with early AD worked with us and were divided into a paper-based art creation therapy group (control group) and a KnowU social teleprompter therapy group for a 6-week intervention. In the digital art creation intervention group we introduced the KnowU digital kit, consisting of a creation plug-in for the Procreate app on a tablet and a wearable device and its app. The entire treatment process is recorded and combined with a quantitative analysis of the McNemar 2 test to analyze the differences in outcomes of verbal communication function in early AD patients after different therapies. Ultimately, it is shown that early AD patients utilizing the KnowU social teleprompter are more effective in the intervention treatment of language decline in the real social domain compared to the paper-based art creation therapy group. The discussion further demonstrates that DTs and art therapy can provide a better social experience, creative approach and emotional recall of language loss in early AD patients, as well as increase the collaborative relationship between early AD patients and their caregivers.
{"title":"KnowU social teleprompter: Interaction design applied to intervention therapy for language decline in early AD patients","authors":"Yi Wu , Minghong Zheng , Changqing Weng","doi":"10.1016/j.visinf.2023.10.004","DOIUrl":"10.1016/j.visinf.2023.10.004","url":null,"abstract":"<div><p>Art therapy as an intervention has been shown to alleviate social impairment in people with AD. Meanwhile, digital technology (DTS) has been shown to perform well in different degenerative dementias through mobile devices and apps. However, it is unclear whether digital art creation therapy has an impact on the speech function of people with early AD. Therefore, the aim of this study was to confirm whether digital art creation therapy has an ameliorating effect on language decline in AD patients through the KnowU social teleprompter. This study was a controlled trial in which 16 patients with early AD worked with us and were divided into a paper-based art creation therapy group (control group) and a KnowU social teleprompter therapy group for a 6-week intervention. In the digital art creation intervention group we introduced the KnowU digital kit, consisting of a creation plug-in for the Procreate app on a tablet and a wearable device and its app. The entire treatment process is recorded and combined with a quantitative analysis of the McNemar <span><math><mi>χ</mi></math></span>2 test to analyze the differences in outcomes of verbal communication function in early AD patients after different therapies. Ultimately, it is shown that early AD patients utilizing the KnowU social teleprompter are more effective in the intervention treatment of language decline in the real social domain compared to the paper-based art creation therapy group. The discussion further demonstrates that DTs and art therapy can provide a better social experience, creative approach and emotional recall of language loss in early AD patients, as well as increase the collaborative relationship between early AD patients and their caregivers.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X23000487/pdfft?md5=35ae9418ed9642f3582f9831393a0416&pid=1-s2.0-S2468502X23000487-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135455325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-01DOI: 10.1016/j.visinf.2022.12.001
Waleed Maqableh , Faisal Y. Alzyoud , Jamal Zraqou
{"title":"Corrigendum to “The Use of Facial Expressions in Measuring Students’ Interaction with Distance Learning Environments During the COVID-19 Crisis” Visual Informatics, Volume 7, Issue 1, March 2023, Pages 1–17","authors":"Waleed Maqableh , Faisal Y. Alzyoud , Jamal Zraqou","doi":"10.1016/j.visinf.2022.12.001","DOIUrl":"10.1016/j.visinf.2022.12.001","url":null,"abstract":"","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":null,"pages":null},"PeriodicalIF":3.0,"publicationDate":"2023-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9747742/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10374441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}