Measuring 3D face deformations from RGB images of expression rehabilitation exercises

Q1 Computer Science Virtual Reality Intelligent Hardware Pub Date : 2022-08-01 DOI:10.1016/j.vrih.2022.05.004
Claudio Ferrari , Stefano Berretti , Pietro Pala , Alberto Del Bimbo
{"title":"Measuring 3D face deformations from RGB images of expression rehabilitation exercises","authors":"Claudio Ferrari ,&nbsp;Stefano Berretti ,&nbsp;Pietro Pala ,&nbsp;Alberto Del Bimbo","doi":"10.1016/j.vrih.2022.05.004","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><p>The accurate (quantitative) analysis of 3D face deformation is a problem of increasing interest in many applications. In particular, defining a 3D model of the face deformation into a 2D target image to capture local and asymmetric deformations remains a challenge in existing literature. A measure of such local deformations may be a relevant index for monitoring the rehabilitation exercises of patients suffering from Parkinson’s or Alzheimer’s disease or those recovering from a stroke.</p></div><div><h3>Methods</h3><p>In this paper, a complete framework that allows the construction of a 3D morphable shape model (3DMM) of the face is presented for fitting to a target RGB image. The model has the specific characteristic of being based on localized components of deformation. The fitting transformation is performed from 3D to 2D and guided by the correspondence between landmarks detected in the target image and those manually annotated on the average 3DMM. The fitting also has the distinction of being performed in two steps to disentangle face deformations related to the identity of the target subject from those induced by facial actions.</p></div><div><h3>Results</h3><p>The method was experimentally validated using the MICC-3D dataset, which includes 11 subjects. Each subject was imaged in one neutral pose and while performing 18 facial actions that deform the face in localized and asymmetric ways. For each acquisition, 3DMM was fit to an RGB frame whereby, from the apex facial action and the neutral frame, the extent of the deformation was computed. The results indicate that the proposed approach can accurately capture face deformation, even localized and asymmetric deformations.</p></div><div><h3>Conclusion</h3><p>The proposed framework demonstrated that it is possible to measure deformations of a reconstructed 3D face model to monitor facial actions performed in response to a set of targets. Interestingly, these results were obtained using only RGB targets, without the need for 3D scans captured with costly devices. This paves the way for the use of the proposed tool in remote medical rehabilitation monitoring.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000456/pdf?md5=10f3974adc62709cdc0d135e68fc356c&pid=1-s2.0-S2096579622000456-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Virtual Reality Intelligent Hardware","FirstCategoryId":"1093","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2096579622000456","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 0

Abstract

Background

The accurate (quantitative) analysis of 3D face deformation is a problem of increasing interest in many applications. In particular, defining a 3D model of the face deformation into a 2D target image to capture local and asymmetric deformations remains a challenge in existing literature. A measure of such local deformations may be a relevant index for monitoring the rehabilitation exercises of patients suffering from Parkinson’s or Alzheimer’s disease or those recovering from a stroke.

Methods

In this paper, a complete framework that allows the construction of a 3D morphable shape model (3DMM) of the face is presented for fitting to a target RGB image. The model has the specific characteristic of being based on localized components of deformation. The fitting transformation is performed from 3D to 2D and guided by the correspondence between landmarks detected in the target image and those manually annotated on the average 3DMM. The fitting also has the distinction of being performed in two steps to disentangle face deformations related to the identity of the target subject from those induced by facial actions.

Results

The method was experimentally validated using the MICC-3D dataset, which includes 11 subjects. Each subject was imaged in one neutral pose and while performing 18 facial actions that deform the face in localized and asymmetric ways. For each acquisition, 3DMM was fit to an RGB frame whereby, from the apex facial action and the neutral frame, the extent of the deformation was computed. The results indicate that the proposed approach can accurately capture face deformation, even localized and asymmetric deformations.

Conclusion

The proposed framework demonstrated that it is possible to measure deformations of a reconstructed 3D face model to monitor facial actions performed in response to a set of targets. Interestingly, these results were obtained using only RGB targets, without the need for 3D scans captured with costly devices. This paves the way for the use of the proposed tool in remote medical rehabilitation monitoring.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用表情康复训练的RGB图像测量三维面部变形
在许多应用中,准确(定量)分析三维人脸变形是一个越来越受关注的问题。特别是,在现有文献中,将面部变形的3D模型定义为2D目标图像以捕获局部和不对称变形仍然是一个挑战。这种局部变形的测量可能是监测帕金森病或阿尔茨海默病患者或中风恢复期患者康复锻炼的相关指标。方法提出了一个完整的人脸三维变形模型(3DMM)构建框架,用于拟合目标RGB图像。该模型具有基于局部变形分量的特点。拟合变换从3D到2D,并根据目标图像中检测到的地标与手动标注在平均3DMM上的地标之间的对应关系进行指导。拟合还具有分两个步骤进行的区别,以将与目标受试者身份相关的面部变形与面部动作引起的面部变形分开。结果采用MICC-3D数据集对该方法进行了实验验证。每位受试者都以一个中立的姿势拍照,同时进行18个面部动作,这些动作会以局部和不对称的方式使面部变形。对于每个采集,3DMM拟合到RGB帧,由此,从顶点面部动作和中性帧,计算变形的程度。结果表明,该方法可以准确地捕获人脸变形,甚至是局部变形和非对称变形。所提出的框架表明,可以测量重建的3D面部模型的变形,以监测面部对一组目标的响应。有趣的是,这些结果仅使用RGB目标获得,而不需要使用昂贵的设备进行3D扫描。这为在远程医疗康复监测中使用拟议的工具铺平了道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Virtual Reality  Intelligent Hardware
Virtual Reality Intelligent Hardware Computer Science-Computer Graphics and Computer-Aided Design
CiteScore
6.40
自引率
0.00%
发文量
35
审稿时长
12 weeks
期刊最新文献
S2ANet: Combining local spectral and spatial point grouping for point cloud processing MKEAH: Multimodal knowledge extraction and accumulation based on hyperplane embedding for knowledge-based visual question answering Generating animatable 3D cartoon faces from single portraits Robust blind image watermarking based on interest points Multi-scale context-aware network for continuous sign language recognition
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1