Pub Date : 2022-03-01DOI: 10.1016/j.visinf.2022.03.002
Yuheng Zhao , Jinjing Jiang , Yi Chen , Richen Liu , Yalong Yang , Xiangyang Xue , Siming Chen
The metaverse is a visual world that blends the physical world and digital world. At present, the development of the metaverse is still in the early stage, and there lacks a framework for the visual construction and exploration of the metaverse. In this paper, we propose a framework that summarizes how graphics, interaction, and visualization techniques support the visual construction of the metaverse and user-centric exploration. We introduce three kinds of visual elements that compose the metaverse and the two graphical construction methods in a pipeline. We propose a taxonomy of interaction technologies based on interaction tasks, user actions, feedback and various sensory channels, and a taxonomy of visualization techniques that assist user awareness. Current potential applications and future opportunities are discussed in the context of visual construction and exploration of the metaverse. We hope this paper can provide a stepping stone for further research in the area of graphics, interaction and visualization in the metaverse.
{"title":"Metaverse: Perspectives from graphics, interactions and visualization","authors":"Yuheng Zhao , Jinjing Jiang , Yi Chen , Richen Liu , Yalong Yang , Xiangyang Xue , Siming Chen","doi":"10.1016/j.visinf.2022.03.002","DOIUrl":"10.1016/j.visinf.2022.03.002","url":null,"abstract":"<div><p>The metaverse is a visual world that blends the physical world and digital world. At present, the development of the metaverse is still in the early stage, and there lacks a framework for the visual construction and exploration of the metaverse. In this paper, we propose a framework that summarizes how graphics, interaction, and visualization techniques support the visual construction of the metaverse and user-centric exploration. We introduce three kinds of visual elements that compose the metaverse and the two graphical construction methods in a pipeline. We propose a taxonomy of interaction technologies based on interaction tasks, user actions, feedback and various sensory channels, and a taxonomy of visualization techniques that assist user awareness. Current potential applications and future opportunities are discussed in the context of visual construction and exploration of the metaverse. We hope this paper can provide a stepping stone for further research in the area of graphics, interaction and visualization in the metaverse.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"6 1","pages":"Pages 56-67"},"PeriodicalIF":3.0,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X22000158/pdfft?md5=1995fb00e264296cfe9e1788841486de&pid=1-s2.0-S2468502X22000158-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122684892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-03-01DOI: 10.1016/j.visinf.2022.01.001
Yongjian Sun , Jie Li , Siming Chen , Gennady Andrienko , Natalia Andrienko , Kang Zhang
We propose an approach to underpin interactive visual exploration of large data volumes by training Learned Visualization Index (LVI). Knowing in advance the data, the aggregation functions that are used for visualization, the visual encoding, and available interactive operations for data selection, LVI allows to avoid time-consuming data retrieval and processing of raw data in response to user’s interactions. Instead, LVI directly predicts aggregates of interest for the user’s data selection. We demonstrate the efficiency of the proposed approach in application to two use cases of spatio-temporal data at different scales.
{"title":"A learning-based approach for efficient visualization construction","authors":"Yongjian Sun , Jie Li , Siming Chen , Gennady Andrienko , Natalia Andrienko , Kang Zhang","doi":"10.1016/j.visinf.2022.01.001","DOIUrl":"10.1016/j.visinf.2022.01.001","url":null,"abstract":"<div><p>We propose an approach to underpin interactive visual exploration of large data volumes by training Learned Visualization Index (LVI). Knowing in advance the data, the aggregation functions that are used for visualization, the visual encoding, and available interactive operations for data selection, LVI allows to avoid time-consuming data retrieval and processing of raw data in response to user’s interactions. Instead, LVI directly predicts aggregates of interest for the user’s data selection. We demonstrate the efficiency of the proposed approach in application to two use cases of spatio-temporal data at different scales.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"6 1","pages":"Pages 14-25"},"PeriodicalIF":3.0,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X22000080/pdfft?md5=16523953bb5f7df328c6c78d0aaff5fa&pid=1-s2.0-S2468502X22000080-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127694326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.12.005
Chengyu Su , Chao Yang , Yonghui Chen , Fupan Wang , Fang Wang , Yadong Wu , Xiaorong Zhang
In the immersive flow visualization based on virtual reality, how to meet the needs of complex professional flow visualization analysis by natural human–computer interaction is a pressing problem. In order to achieve the natural and efficient human–computer interaction, we analyze the interaction requirements of flow visualization and study the characteristics of four human–computer interaction channels: hand, head, eye and voice. We give out some multimodal interaction design suggestions and then propose three multimodal interaction methods: head & hand, head & hand & eye and head & hand & eye & voice. The freedom of gestures, the stability of the head, the convenience of eyes and the rapid retrieval of voices are used to improve the accuracy and efficiency of interaction. The interaction load is balanced by multimodal interaction to reduce fatigue. The evaluation shows that our multimodal interaction has higher accuracy, faster time efficiency and much lower fatigue than the traditional joystick interaction.
{"title":"Natural multimodal interaction in immersive flow visualization","authors":"Chengyu Su , Chao Yang , Yonghui Chen , Fupan Wang , Fang Wang , Yadong Wu , Xiaorong Zhang","doi":"10.1016/j.visinf.2021.12.005","DOIUrl":"10.1016/j.visinf.2021.12.005","url":null,"abstract":"<div><p>In the immersive flow visualization based on virtual reality, how to meet the needs of complex professional flow visualization analysis by natural human–computer interaction is a pressing problem. In order to achieve the natural and efficient human–computer interaction, we analyze the interaction requirements of flow visualization and study the characteristics of four human–computer interaction channels: hand, head, eye and voice. We give out some multimodal interaction design suggestions and then propose three multimodal interaction methods: head & hand, head & hand & eye and head & hand & eye & voice. The freedom of gestures, the stability of the head, the convenience of eyes and the rapid retrieval of voices are used to improve the accuracy and efficiency of interaction. The interaction load is balanced by multimodal interaction to reduce fatigue. The evaluation shows that our multimodal interaction has higher accuracy, faster time efficiency and much lower fatigue than the traditional joystick interaction.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 56-66"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000632/pdfft?md5=28eaf01c7c9f0094b805091c154c6864&pid=1-s2.0-S2468502X21000632-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133259156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.12.002
Wei Zhang , Qian Ma , Rusheng Pan , Wei Chen
Song Ci is treasured in traditional Chinese culture, which indicates social and cultural evolution in ancient times. Despite the efforts by historians and litterateurs in investigating the characteristics of Song Ci, it is still unclear how to effectively distribute and promote Song Ci in the public sphere. The complexity and abstraction of Song Ci hamper the general public from closely reading, analyzing, and appreciating these excellent works. By means of a set of visual analysis methods, e.g. the spatio-temporal visualization, we exploit visual storytelling to explicitly present the latent and abstractive features of Song Ci. We apply straightway visual charts and lighten the burden of understanding the stories, in order to achieve an effective public distribution. The effectiveness and aesthetics of our work are demonstrated by a user study of three participants with different backgrounds. The result reveals that our story is effective in the distribution, understanding, and promotion of Song Ci.
{"title":"Visual storytelling of Song Ci and the poets in the social–cultural context of Song dynasty","authors":"Wei Zhang , Qian Ma , Rusheng Pan , Wei Chen","doi":"10.1016/j.visinf.2021.12.002","DOIUrl":"10.1016/j.visinf.2021.12.002","url":null,"abstract":"<div><p>Song Ci is treasured in traditional Chinese culture, which indicates social and cultural evolution in ancient times. Despite the efforts by historians and litterateurs in investigating the characteristics of Song Ci, it is still unclear how to effectively distribute and promote Song Ci in the public sphere. The complexity and abstraction of Song Ci hamper the general public from closely reading, analyzing, and appreciating these excellent works. By means of a set of visual analysis methods, e.g. the spatio-temporal visualization, we exploit visual storytelling to explicitly present the latent and abstractive features of Song Ci. We apply straightway visual charts and lighten the burden of understanding the stories, in order to achieve an effective public distribution. The effectiveness and aesthetics of our work are demonstrated by a user study of three participants with different backgrounds. The result reveals that our story is effective in the distribution, understanding, and promotion of Song Ci.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 34-40"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000607/pdfft?md5=7de35aa21498ec269f63ff4c547c971e&pid=1-s2.0-S2468502X21000607-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121839735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.12.001
Sukanta Ghosh, Amlan Pratim Hazarika, Abhijit Chandra, Rajani K. Mudi
Progression of Alzheimer’s disease (AD) bears close proximity with the tissue loss in the medial temporal lobe (MTL) and enlargement of lateral ventricle (LV). The early stage of AD, mild cognitive impairment (MCI), can be traced by diagnosing brain MRI scans with advanced fuzzy c-means clustering algorithm that helps to take an appropriate intervention. In this paper, firstly the sparsity is initiated in clustering method that too rician noise is also incorporated for brain MR scans of AD subject. Secondly, a novel neighbor pixel constrained fuzzy c-means clustering algorithm is designed where topoloty-based selection of parsimonious neighbor pixels is automated. The adaptability in choice of neighbor pixel class outliers more justified object edge boundary which outperforms a dynamic cluster output. The proposed adaptive neighbor constrained deviation sparse variant fuzzy c-means clustering (AN_DsFCM) can withhold imposed sparsity and withstands rician noise at imposed sparse environment. This novel algorithm is applied for MRI of AD subjects and normative data is acquired to analyse clustering accuracy. The data processing pipeline of theoretically plausible proposition is elaborated in detail. The experimental results are compared with state-of-the-art fuzzy clustering methods for test MRI scans. Visual evaluation and statistical measures are studied to meet both image processing and clinical neurophysiology standards. Overall the performance of proposed AN_DsFCM is significantly better than other methods.
{"title":"Adaptive neighbor constrained deviation sparse variant fuzzy c-means clustering for brain MRI of AD subject","authors":"Sukanta Ghosh, Amlan Pratim Hazarika, Abhijit Chandra, Rajani K. Mudi","doi":"10.1016/j.visinf.2021.12.001","DOIUrl":"10.1016/j.visinf.2021.12.001","url":null,"abstract":"<div><p>Progression of Alzheimer’s disease (AD) bears close proximity with the tissue loss in the medial temporal lobe (MTL) and enlargement of lateral ventricle (LV). The early stage of AD, mild cognitive impairment (MCI), can be traced by diagnosing brain MRI scans with advanced fuzzy c-means clustering algorithm that helps to take an appropriate intervention. In this paper, firstly the sparsity is initiated in clustering method that too rician noise is also incorporated for brain MR scans of AD subject. Secondly, a novel neighbor pixel constrained fuzzy c-means clustering algorithm is designed where topoloty-based selection of parsimonious neighbor pixels is automated. The adaptability in choice of neighbor pixel class outliers more justified object edge boundary which outperforms a dynamic cluster output. The proposed adaptive neighbor constrained deviation sparse variant fuzzy c-means clustering (AN_DsFCM) can withhold imposed sparsity and withstands rician noise at imposed sparse environment. This novel algorithm is applied for MRI of AD subjects and normative data is acquired to analyse clustering accuracy. The data processing pipeline of theoretically plausible proposition is elaborated in detail. The experimental results are compared with state-of-the-art fuzzy clustering methods for test MRI scans. Visual evaluation and statistical measures are studied to meet both image processing and clinical neurophysiology standards. Overall the performance of proposed AN_DsFCM is significantly better than other methods.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 67-80"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000589/pdfft?md5=129443dd9a0b4930d3d376ba5efb1317&pid=1-s2.0-S2468502X21000589-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128584891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.11.001
Zhipeng Hu , Changjie Fan , Qiwei Zheng , Wei Wu , Bai Liu
Visual programming tools are widely applied in the game industry to assist game designers in developing game artificial intelligence (game AI) and gameplay. However, testing multiple game engines is a time-consuming operation, which degrades development efficiency. To provide an asynchronous platform for game designers, this paper introduces Asyncflow, an open-source visual programming solution. It consists of a flowchart maker for game logic explanation and a runtime framework integrating an asynchronous mechanism based on an event-driven architecture. Asyncflow supports multiple programming languages and can be easily embedded in various game engines to run flowcharts created by game designers.
{"title":"Asyncflow: A visual programming tool for game artificial intelligence","authors":"Zhipeng Hu , Changjie Fan , Qiwei Zheng , Wei Wu , Bai Liu","doi":"10.1016/j.visinf.2021.11.001","DOIUrl":"10.1016/j.visinf.2021.11.001","url":null,"abstract":"<div><p>Visual programming tools are widely applied in the game industry to assist game designers in developing game artificial intelligence (game AI) and gameplay. However, testing multiple game engines is a time-consuming operation, which degrades development efficiency. To provide an asynchronous platform for game designers, this paper introduces <em>Asyncflow</em>, an open-source visual programming solution. It consists of a flowchart maker for game logic explanation and a runtime framework integrating an asynchronous mechanism based on an event-driven architecture. <em>Asyncflow</em> supports multiple programming languages and can be easily embedded in various game engines to run flowcharts created by game designers.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 20-25"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000498/pdfft?md5=665ce823bd9a7b3c5b3dc285847edd6f&pid=1-s2.0-S2468502X21000498-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125256391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.10.002
Haiyan Liu , Xiaohui Chen , Yidi Wang , Bing Zhang , Yunpeng Chen , Ying Zhao , Fangfang Zhou
Maritime transports play a critical role in international trade and commerce. Massive vessels sailing around the world continuously generate vessel trajectory data that contain rich spatial–temporal patterns of vessel navigations. Analyzing and understanding these patterns are valuable for maritime traffic surveillance and management. As essential techniques in complex data analysis and understanding, visualization and visual analysis have been widely used in vessel trajectory data analysis. This paper presents a literature review on the visualization and visual analysis of vessel trajectory data. First, we introduce commonly used vessel trajectory data sets and summarize main operations in vessel trajectory data preprocessing. Then, we provide a taxonomy of visualization and visual analysis of vessel trajectory data based on existing approaches and introduce representative works in details. Finally, we expound on the prospects of the remaining challenges and directions for future research.
{"title":"Visualization and visual analysis of vessel trajectory data: A survey","authors":"Haiyan Liu , Xiaohui Chen , Yidi Wang , Bing Zhang , Yunpeng Chen , Ying Zhao , Fangfang Zhou","doi":"10.1016/j.visinf.2021.10.002","DOIUrl":"10.1016/j.visinf.2021.10.002","url":null,"abstract":"<div><p>Maritime transports play a critical role in international trade and commerce. Massive vessels sailing around the world continuously generate vessel trajectory data that contain rich spatial–temporal patterns of vessel navigations. Analyzing and understanding these patterns are valuable for maritime traffic surveillance and management. As essential techniques in complex data analysis and understanding, visualization and visual analysis have been widely used in vessel trajectory data analysis. This paper presents a literature review on the visualization and visual analysis of vessel trajectory data. First, we introduce commonly used vessel trajectory data sets and summarize main operations in vessel trajectory data preprocessing. Then, we provide a taxonomy of visualization and visual analysis of vessel trajectory data based on existing approaches and introduce representative works in details. Finally, we expound on the prospects of the remaining challenges and directions for future research.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 1-10"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000401/pdfft?md5=c0cbcf1e197c069abdaada9d25a133d0&pid=1-s2.0-S2468502X21000401-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127063228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Edges crossing and nodes overlapping have a significant effect on the users’ recognition and comprehension of network diagrams. In this study, we propose a visual evaluation method for users’ cognition of network diagrams. First, this method carries out a set of cognitive experiments to collect the user’s cognitive performance that affects the variables, including accuracy and response time. The user’s pupil diameter is measured through an eye tracker to reflect their cognitive load. Second, the significance test points out the visual features as independent variables and then establishes an evaluation regression model. The experimental results show that the number of edges, edge length, node visual interference, and edge occlusion contribute to the evaluation models of response time, and edge occlusion and the number of node connections contribute to the accuracy model. Finally, these evaluation models demonstrate good predictability when assessing users’ cognition of network diagrams and provide practical recommendations for their use.
{"title":"Evaluating user cognition of network diagrams","authors":"Xiaojiao Chen , Xiaoteng Tang , Zijing Luo , Jiayi Zhang","doi":"10.1016/j.visinf.2021.12.004","DOIUrl":"10.1016/j.visinf.2021.12.004","url":null,"abstract":"<div><p>Edges crossing and nodes overlapping have a significant effect on the users’ recognition and comprehension of network diagrams. In this study, we propose a visual evaluation method for users’ cognition of network diagrams. First, this method carries out a set of cognitive experiments to collect the user’s cognitive performance that affects the variables, including accuracy and response time. The user’s pupil diameter is measured through an eye tracker to reflect their cognitive load. Second, the significance test points out the visual features as independent variables and then establishes an evaluation regression model. The experimental results show that the number of edges, edge length, node visual interference, and edge occlusion contribute to the evaluation models of response time, and edge occlusion and the number of node connections contribute to the accuracy model. Finally, these evaluation models demonstrate good predictability when assessing users’ cognition of network diagrams and provide practical recommendations for their use.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 26-33"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000620/pdfft?md5=cb42e16ec678484dc019ff590291645b&pid=1-s2.0-S2468502X21000620-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132205860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.12.003
Yanyan Wang , Zhanning Bai , Zhifeng Lin , Xiaoqing Dong , Yingchaojie Feng , Jiacheng Pan , Wei Chen
Authoring graph visualization poses great challenges to developers due to its high requirements on both domain knowledge and development skills. Although existing libraries and tools reduce the difficulty of generating graph visualization, there are still many challenges. We work closely with developers and formulate several design goals, then design and implement G6, a web-based library for graph visualization. It combines template-based configuration for high usability and flexible customization for high expressiveness. To enhance development efficiency, G6 proposes a range of optimizations, including state management and interaction modes. We demonstrate its capabilities through an extensive gallery, a quantitative performance evaluation, and an expert interview. G6 was first released in 2017 and has been iterated for 317 versions. It has served as a web-based library for thousands of applications and received 8312 stars on GitHub.
{"title":"G6: A web-based library for graph visualization","authors":"Yanyan Wang , Zhanning Bai , Zhifeng Lin , Xiaoqing Dong , Yingchaojie Feng , Jiacheng Pan , Wei Chen","doi":"10.1016/j.visinf.2021.12.003","DOIUrl":"10.1016/j.visinf.2021.12.003","url":null,"abstract":"<div><p>Authoring graph visualization poses great challenges to developers due to its high requirements on both domain knowledge and development skills. Although existing libraries and tools reduce the difficulty of generating graph visualization, there are still many challenges. We work closely with developers and formulate several design goals, then design and implement G6, a web-based library for graph visualization. It combines template-based configuration for high usability and flexible customization for high expressiveness. To enhance development efficiency, G6 proposes a range of optimizations, including state management and interaction modes. We demonstrate its capabilities through an extensive gallery, a quantitative performance evaluation, and an expert interview. G6 was first released in 2017 and has been iterated for 317 versions. It has served as a web-based library for thousands of applications and received 8312 stars on GitHub.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 49-55"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000619/pdfft?md5=a1d76898fda1924291672cf6aef0191a&pid=1-s2.0-S2468502X21000619-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128521952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-01DOI: 10.1016/j.visinf.2021.11.003
Jiesi Li , Ning Xu , Weizhi Nie , Shenyuan Zhang
Image Captioning is a cross-modal task that needs to automatically generate coherent natural sentences to describe the image contents. Due to the large gap between vision and language modalities, most of the existing methods have the problem of inaccurate semantic matching between images and generated captions. To solve the problem, this paper proposes a novel multi-level similarity-guided semantic matching method for image captioning, which can fuse local and global semantic similarities to learn the latent semantic correlation between images and generated captions. Specifically, we extract the semantic units containing fine-grained semantic information of images and generated captions, respectively. Based on the comparison of the semantic units, we design a local semantic similarity evaluation mechanism. Meanwhile, we employ the CIDEr score to characterize the global semantic similarity. The local and global two-level similarities are finally fused using the reinforcement learning theory, to guide the model optimization to obtain better semantic matching. The quantitative and qualitative experiments on large-scale MSCOCO dataset illustrate the superiority of the proposed method, which can achieve fine-grained semantic matching of images and generated captions.
{"title":"Image Captioning with multi-level similarity-guided semantic matching","authors":"Jiesi Li , Ning Xu , Weizhi Nie , Shenyuan Zhang","doi":"10.1016/j.visinf.2021.11.003","DOIUrl":"10.1016/j.visinf.2021.11.003","url":null,"abstract":"<div><p>Image Captioning is a cross-modal task that needs to automatically generate coherent natural sentences to describe the image contents. Due to the large gap between vision and language modalities, most of the existing methods have the problem of inaccurate semantic matching between images and generated captions. To solve the problem, this paper proposes a novel multi-level similarity-guided semantic matching method for image captioning, which can fuse local and global semantic similarities to learn the latent semantic correlation between images and generated captions. Specifically, we extract the semantic units containing fine-grained semantic information of images and generated captions, respectively. Based on the comparison of the semantic units, we design a local semantic similarity evaluation mechanism. Meanwhile, we employ the CIDEr score to characterize the global semantic similarity. The local and global two-level similarities are finally fused using the reinforcement learning theory, to guide the model optimization to obtain better semantic matching. The quantitative and qualitative experiments on large-scale MSCOCO dataset illustrate the superiority of the proposed method, which can achieve fine-grained semantic matching of images and generated captions.</p></div>","PeriodicalId":36903,"journal":{"name":"Visual Informatics","volume":"5 4","pages":"Pages 41-48"},"PeriodicalIF":3.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2468502X21000590/pdfft?md5=f944bc3d86f6d64595ece2bbaa4a94c8&pid=1-s2.0-S2468502X21000590-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124498801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}