Consistent Long Sequences Deep Faces

Xudong Fan, Daniele Bonatto, G. Lafruit
{"title":"Consistent Long Sequences Deep Faces","authors":"Xudong Fan, Daniele Bonatto, G. Lafruit","doi":"10.1109/IC3D48390.2019.8975999","DOIUrl":null,"url":null,"abstract":"Face swapping in videos usually has strong entertainment applications. Deep Fakes (in Faces) are a recent topic in deep learning where the main idea is to substitute the face of a person in a video with the face of another person. But one of the drawbacks of the method is that between two successive frames there are inconsistencies between the faces, such as changing face color, flickering or eyebrows that change. In this paper, we propose a convolutional neural network for swapping faces based on two autoencoders which share the same encoder. In this network, the encoder can distinguish and extract important features of faces, including facial expressions and poses; the decoders will then reconstruct faces according to these features. First, we will generate datasets of faces respectively for person A and person B. Secondly, the local information of two faces is sent to the network to get the model; after the training process, we can use the model to reconstruct the corresponding face of person B when the input is one face of person A. Afterwards, we build a binary mask to select the face area and transfer color from the source face to the target face. Finally, we only need to use a seamless clone to merge the new faces back into the source frames to create a fake video. The experimental results show that the quality of the fake videos is improved significantly.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"130 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on 3D Immersion (IC3D)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IC3D48390.2019.8975999","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Face swapping in videos usually has strong entertainment applications. Deep Fakes (in Faces) are a recent topic in deep learning where the main idea is to substitute the face of a person in a video with the face of another person. But one of the drawbacks of the method is that between two successive frames there are inconsistencies between the faces, such as changing face color, flickering or eyebrows that change. In this paper, we propose a convolutional neural network for swapping faces based on two autoencoders which share the same encoder. In this network, the encoder can distinguish and extract important features of faces, including facial expressions and poses; the decoders will then reconstruct faces according to these features. First, we will generate datasets of faces respectively for person A and person B. Secondly, the local information of two faces is sent to the network to get the model; after the training process, we can use the model to reconstruct the corresponding face of person B when the input is one face of person A. Afterwards, we build a binary mask to select the face area and transfer color from the source face to the target face. Finally, we only need to use a seamless clone to merge the new faces back into the source frames to create a fake video. The experimental results show that the quality of the fake videos is improved significantly.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一致的长序列深面
视频中的人脸交换通常具有很强的娱乐应用。Deep Fakes (in Faces)是深度学习领域最近的一个话题,其主要思想是用另一个人的脸代替视频中的人的脸。但该方法的缺点之一是,在两个连续的帧之间,面部之间存在不一致,例如面部颜色的变化,闪烁或眉毛的变化。在本文中,我们提出了一种基于两个共享同一编码器的自编码器交换人脸的卷积神经网络。在该网络中,编码器可以区分和提取人脸的重要特征,包括面部表情和姿势;然后,解码器将根据这些特征重建人脸。首先,我们将分别生成A人和b人的人脸数据集,然后将两张人脸的局部信息发送到网络中得到模型;训练过程结束后,当输入是a人的一张脸时,我们可以使用该模型重构出B人对应的脸。然后,我们构建一个二值掩码来选择人脸区域,并将源人脸的颜色转移到目标人脸。最后,我们只需要使用无缝克隆将新面孔合并回源帧以创建假视频。实验结果表明,伪视频的质量得到了显著提高。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Consistent Long Sequences Deep Faces A Novel Randomize Hierarchical Extension of MV-HEVC for Improved Light Field Compression A Novel Algebaric Variety Based Model for High Quality Free-Viewpoint View Synthesis on a Krylov Subspace Relating Eye Dominance to Neurochemistry in the Human Visual Cortex Using Ultra High Field 7-Tesla MR Spectroscopy Frame-Wise CNN-Based View Synthesis for Light Field Camera Arrays
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1