EventPointMesh: Human Mesh Recovery Solely From Event Point Clouds

Ryosuke Hori;Mariko Isogawa;Dan Mikami;Hideo Saito
{"title":"EventPointMesh: Human Mesh Recovery Solely From Event Point Clouds","authors":"Ryosuke Hori;Mariko Isogawa;Dan Mikami;Hideo Saito","doi":"10.1109/TVCG.2024.3462816","DOIUrl":null,"url":null,"abstract":"How much can we infer about human shape using an event camera that only detects the pixel position where the luminance changed and its timestamp? This neuromorphic vision technology captures changes in pixel values at ultra-high speeds, regardless of the variations in environmental lighting brightness. Existing methods for human mesh recovery (HMR) from event data need to utilize intensity images captured with a generic frame-based camera, rendering them vulnerable to low-light conditions, energy/memory constraints, and privacy issues. In contrast, we explore the potential of solely utilizing event data to alleviate these issues and ascertain whether it offers adequate cues for HMR, as illustrated in Fig. 1. This is a quite challenging task due to the substantially limited information ensuing from the absence of intensity images. To this end, we propose EventPointMesh, a framework which treats event data as a three-dimensional (3D) spatio-temporal point cloud for reconstructing the human mesh. By employing a coarse-to-fine pose feature extraction strategy, we extract both global features and local features. The local features are derived by processing the spatio-temporally dispersed event points into groups associated with individual body segments. This combination of global and local features allows the framework to achieve a more accurate HMR, capturing subtle differences in human movements. Experiments demonstrate that our method with only sparse event data outperforms baseline methods.","PeriodicalId":94035,"journal":{"name":"IEEE transactions on visualization and computer graphics","volume":"31 9","pages":"5593-5610"},"PeriodicalIF":6.5000,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10683888","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on visualization and computer graphics","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10683888/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

How much can we infer about human shape using an event camera that only detects the pixel position where the luminance changed and its timestamp? This neuromorphic vision technology captures changes in pixel values at ultra-high speeds, regardless of the variations in environmental lighting brightness. Existing methods for human mesh recovery (HMR) from event data need to utilize intensity images captured with a generic frame-based camera, rendering them vulnerable to low-light conditions, energy/memory constraints, and privacy issues. In contrast, we explore the potential of solely utilizing event data to alleviate these issues and ascertain whether it offers adequate cues for HMR, as illustrated in Fig. 1. This is a quite challenging task due to the substantially limited information ensuing from the absence of intensity images. To this end, we propose EventPointMesh, a framework which treats event data as a three-dimensional (3D) spatio-temporal point cloud for reconstructing the human mesh. By employing a coarse-to-fine pose feature extraction strategy, we extract both global features and local features. The local features are derived by processing the spatio-temporally dispersed event points into groups associated with individual body segments. This combination of global and local features allows the framework to achieve a more accurate HMR, capturing subtle differences in human movements. Experiments demonstrate that our method with only sparse event data outperforms baseline methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
事件点网格:仅从事件点云恢复人体网格
使用仅检测亮度变化的像素位置及其时间戳的事件相机,我们能推断出多少关于人体形状的信息?这种神经形态视觉技术以超高速捕捉像素值的变化,而不考虑环境照明亮度的变化。现有的从事件数据中进行人体网格恢复(HMR)的方法需要利用一般基于帧的相机捕获的强度图像,这使得它们容易受到低光照条件、能量/内存限制和隐私问题的影响。相反,我们探索了单独利用事件数据来缓解这些问题的潜力,并确定它是否为HMR提供了足够的线索,如图1所示。这是一项相当具有挑战性的任务,因为缺乏强度图像所带来的信息非常有限。为此,我们提出了EventPointMesh框架,该框架将事件数据作为三维(3D)时空点云来重建人体网格。采用从粗到精的姿态特征提取策略,提取全局特征和局部特征。局部特征是通过将时空分散的事件点处理成与个体体段相关的组而得到的。这种整体和局部特征的结合使框架能够实现更准确的HMR,捕捉人类运动的细微差异。实验表明,我们的方法仅使用稀疏事件数据就优于基线方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
ESGaussianFace: Emotional and Stylized Audio-Driven Facial Animation Via 3D Gaussian Splatting. Generating Distance-Aware Human-to-Human Interaction Motions From Text Guidance. LAMDA: Aiding Visual Exploration of Atomic Displacements in Molecular Dynamics Simulations. Locally Adapted Reference Frame Fields using Moving Least Squares. SeparateGen: Semantic Component-based 3D Character Generation from Single Images.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1