Yudi Dai, Zhiyong Wang, Xiping Lin, Chenglu Wen, Lan Xu, Siqi Shen, Yuexin Ma, Cheng Wang
{"title":"HiSC4D: Human-centered interaction and 4D Scene Capture in Large-scale Space Using Wearable IMUs and LiDAR","authors":"Yudi Dai, Zhiyong Wang, Xiping Lin, Chenglu Wen, Lan Xu, Siqi Shen, Yuexin Ma, Cheng Wang","doi":"arxiv-2409.04398","DOIUrl":null,"url":null,"abstract":"We introduce HiSC4D, a novel Human-centered interaction and 4D Scene Capture\nmethod, aimed at accurately and efficiently creating a dynamic digital world,\ncontaining large-scale indoor-outdoor scenes, diverse human motions, rich\nhuman-human interactions, and human-environment interactions. By utilizing\nbody-mounted IMUs and a head-mounted LiDAR, HiSC4D can capture egocentric human\nmotions in unconstrained space without the need for external devices and\npre-built maps. This affords great flexibility and accessibility for\nhuman-centered interaction and 4D scene capturing in various environments.\nTaking into account that IMUs can capture human spatially unrestricted poses\nbut are prone to drifting for long-period using, and while LiDAR is stable for\nglobal localization but rough for local positions and orientations, HiSC4D\nemploys a joint optimization method, harmonizing all sensors and utilizing\nenvironment cues, yielding promising results for long-term capture in large\nscenes. To promote research of egocentric human interaction in large scenes and\nfacilitate downstream tasks, we also present a dataset, containing 8 sequences\nin 4 large scenes (200 to 5,000 $m^2$), providing 36k frames of accurate 4D\nhuman motions with SMPL annotations and dynamic scenes, 31k frames of cropped\nhuman point clouds, and scene mesh of the environment. A variety of scenarios,\nsuch as the basketball gym and commercial street, alongside challenging human\nmotions, such as daily greeting, one-on-one basketball playing, and tour\nguiding, demonstrate the effectiveness and the generalization ability of\nHiSC4D. The dataset and code will be publicated on\nwww.lidarhumanmotion.net/hisc4d available for research purposes.","PeriodicalId":501480,"journal":{"name":"arXiv - CS - Multimedia","volume":"34 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.04398","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We introduce HiSC4D, a novel Human-centered interaction and 4D Scene Capture
method, aimed at accurately and efficiently creating a dynamic digital world,
containing large-scale indoor-outdoor scenes, diverse human motions, rich
human-human interactions, and human-environment interactions. By utilizing
body-mounted IMUs and a head-mounted LiDAR, HiSC4D can capture egocentric human
motions in unconstrained space without the need for external devices and
pre-built maps. This affords great flexibility and accessibility for
human-centered interaction and 4D scene capturing in various environments.
Taking into account that IMUs can capture human spatially unrestricted poses
but are prone to drifting for long-period using, and while LiDAR is stable for
global localization but rough for local positions and orientations, HiSC4D
employs a joint optimization method, harmonizing all sensors and utilizing
environment cues, yielding promising results for long-term capture in large
scenes. To promote research of egocentric human interaction in large scenes and
facilitate downstream tasks, we also present a dataset, containing 8 sequences
in 4 large scenes (200 to 5,000 $m^2$), providing 36k frames of accurate 4D
human motions with SMPL annotations and dynamic scenes, 31k frames of cropped
human point clouds, and scene mesh of the environment. A variety of scenarios,
such as the basketball gym and commercial street, alongside challenging human
motions, such as daily greeting, one-on-one basketball playing, and tour
guiding, demonstrate the effectiveness and the generalization ability of
HiSC4D. The dataset and code will be publicated on
www.lidarhumanmotion.net/hisc4d available for research purposes.