Peiliang Wu, Haozhe Zhang, Yao Li, Wenbai Chen, Guowei Gao
{"title":"VT-SGN:Spiking Graph Neural Network for Neuromorphic Visual-Tactile Fusion.","authors":"Peiliang Wu, Haozhe Zhang, Yao Li, Wenbai Chen, Guowei Gao","doi":"10.1109/TOH.2024.3449411","DOIUrl":null,"url":null,"abstract":"<p><p>Current issues with neuromorphic visual-tactile perception include limited training network representation and inadequate cross-modal fusion. To address these two issues, we proposed a dual network called visual-tactile spiking graph neural network (VT-SGN) that combines graph neural networks and spiking neural networks to jointly utilize the neuromorphic visual and tactile source data. First, the neuromorphic visual-tactile data were expanded spatiotemporally to create a taxel-based tactile graph in the spatial domain, enabling the complete exploitation of the irregular spatial structure properties of tactile information. Subsequently, a method for converting images into graph structures was proposed, allowing the vision to be trained alongside graph neural networks and extracting graph-level features from the vision for fusion with tactile data. Finally, the data were expanded into the time domain using a spiking neural network to train the model and propagate it backwards. This framework effectively utilizes the structural differences between sample instances in the spatial dimension to improve the representational power of spiking neurons, while preserving the biodynamic mechanism of the spiking neural network. Additionally, it effectively solves the morphological variance between the two perceptions and further uses complementary data between visual and tactile. To demonstrate that our approach can improve the learning of neuromorphic perceptual information, we conducted comprehensive comparative experiments on three datasets to validate the benefits of the proposed VT-SGN framework by comparing it with state-of-the-art studies.</p>","PeriodicalId":13215,"journal":{"name":"IEEE Transactions on Haptics","volume":"PP ","pages":""},"PeriodicalIF":2.4000,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Haptics","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1109/TOH.2024.3449411","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
引用次数: 0
Abstract
Current issues with neuromorphic visual-tactile perception include limited training network representation and inadequate cross-modal fusion. To address these two issues, we proposed a dual network called visual-tactile spiking graph neural network (VT-SGN) that combines graph neural networks and spiking neural networks to jointly utilize the neuromorphic visual and tactile source data. First, the neuromorphic visual-tactile data were expanded spatiotemporally to create a taxel-based tactile graph in the spatial domain, enabling the complete exploitation of the irregular spatial structure properties of tactile information. Subsequently, a method for converting images into graph structures was proposed, allowing the vision to be trained alongside graph neural networks and extracting graph-level features from the vision for fusion with tactile data. Finally, the data were expanded into the time domain using a spiking neural network to train the model and propagate it backwards. This framework effectively utilizes the structural differences between sample instances in the spatial dimension to improve the representational power of spiking neurons, while preserving the biodynamic mechanism of the spiking neural network. Additionally, it effectively solves the morphological variance between the two perceptions and further uses complementary data between visual and tactile. To demonstrate that our approach can improve the learning of neuromorphic perceptual information, we conducted comprehensive comparative experiments on three datasets to validate the benefits of the proposed VT-SGN framework by comparing it with state-of-the-art studies.
期刊介绍:
IEEE Transactions on Haptics (ToH) is a scholarly archival journal that addresses the science, technology, and applications associated with information acquisition and object manipulation through touch. Haptic interactions relevant to this journal include all aspects of manual exploration and manipulation of objects by humans, machines and interactions between the two, performed in real, virtual, teleoperated or networked environments. Research areas of relevance to this publication include, but are not limited to, the following topics: Human haptic and multi-sensory perception and action, Aspects of motor control that explicitly pertain to human haptics, Haptic interactions via passive or active tools and machines, Devices that sense, enable, or create haptic interactions locally or at a distance, Haptic rendering and its association with graphic and auditory rendering in virtual reality, Algorithms, controls, and dynamics of haptic devices, users, and interactions between the two, Human-machine performance and safety with haptic feedback, Haptics in the context of human-computer interactions, Systems and networks using haptic devices and interactions, including multi-modal feedback, Application of the above, for example in areas such as education, rehabilitation, medicine, computer-aided design, skills training, computer games, driver controls, simulation, and visualization.