SPADNet: Structure Prior-Aware Dynamic Network for Face Super-Resolution

Chenyang Wang;Junjun Jiang;Kui Jiang;Xianming Liu
{"title":"SPADNet: Structure Prior-Aware Dynamic Network for Face Super-Resolution","authors":"Chenyang Wang;Junjun Jiang;Kui Jiang;Xianming Liu","doi":"10.1109/TBIOM.2024.3382870","DOIUrl":null,"url":null,"abstract":"The recent emergence of deep learning neural networks has propelled advancements in the field of face super-resolution. While these deep learning-based methods have shown significant performance improvements, they depend overwhelmingly on fixed, spatially shared kernels within standard convolutional layers. This leads to a neglect of the diverse facial structures and regions, consequently struggling to reconstruct high-fidelity face images. As a highly structured object, the structural features of a face are crucial for representing and reconstructing face images. To this end, we introduce a structure prior-aware dynamic network (SPADNet) that leverages facial structure priors as a foundation to generate structure-aware dynamic kernels for the distinctive super-resolution of various face images. In view of that spatially shared kernels are not well-suited for specific-regions representation, a local structure-adaptive convolution (LSAC) is devised to characterize the local relation of facial features. It is more effective for precise texture representation. Meanwhile, a global structure-aware convolution (GSAC) is elaborated to capture the global facial contours to guarantee the structure consistency. These strategies form a unified face reconstruction framework, which reconciles the distinct representation of diverse face images and individual structure fidelity. Extensive experiments confirm the superiority of our proposed SPADNet over state-of-the-art methods. The source codes of the proposed method will be available at \n<uri>https://github.com/wcy-cs/SPADNet</uri>\n.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 3","pages":"326-340"},"PeriodicalIF":5.0000,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on biometrics, behavior, and identity science","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10485196/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The recent emergence of deep learning neural networks has propelled advancements in the field of face super-resolution. While these deep learning-based methods have shown significant performance improvements, they depend overwhelmingly on fixed, spatially shared kernels within standard convolutional layers. This leads to a neglect of the diverse facial structures and regions, consequently struggling to reconstruct high-fidelity face images. As a highly structured object, the structural features of a face are crucial for representing and reconstructing face images. To this end, we introduce a structure prior-aware dynamic network (SPADNet) that leverages facial structure priors as a foundation to generate structure-aware dynamic kernels for the distinctive super-resolution of various face images. In view of that spatially shared kernels are not well-suited for specific-regions representation, a local structure-adaptive convolution (LSAC) is devised to characterize the local relation of facial features. It is more effective for precise texture representation. Meanwhile, a global structure-aware convolution (GSAC) is elaborated to capture the global facial contours to guarantee the structure consistency. These strategies form a unified face reconstruction framework, which reconciles the distinct representation of diverse face images and individual structure fidelity. Extensive experiments confirm the superiority of our proposed SPADNet over state-of-the-art methods. The source codes of the proposed method will be available at https://github.com/wcy-cs/SPADNet .
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
SPADNet:用于人脸超分辨率的结构先验感知动态网络
最近出现的深度学习神经网络推动了人脸超分辨率领域的进步。虽然这些基于深度学习的方法在性能上有了显著提高,但它们绝大多数都依赖于标准卷积层中固定的、空间共享的内核。这导致了对不同面部结构和区域的忽视,从而难以重建高保真的面部图像。作为一个高度结构化的对象,人脸的结构特征对于表现和重建人脸图像至关重要。为此,我们引入了一种结构先验感知动态网络(SPADNet),该网络以人脸结构先验为基础,生成结构感知动态核,用于对各种人脸图像进行独特的超分辨率处理。鉴于空间共享内核并不适合特定区域的表示,因此设计了局部结构自适应卷积(LSAC)来表征面部特征的局部关系。它对精确的纹理表示更为有效。同时,全局结构感知卷积(GSAC)用于捕捉全局面部轮廓,以保证结构的一致性。这些策略构成了一个统一的人脸重建框架,它兼顾了不同人脸图像的不同表示和个体结构的保真度。大量实验证实,我们提出的 SPADNet 优于最先进的方法。建议方法的源代码将公布在 https://github.com/wcy-cs/SPADNet 网站上。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
10.90
自引率
0.00%
发文量
0
期刊最新文献
Table of Contents Editorial for the TBIOM Special Issue on Generative AI and Large Vision-Language Models for Biometrics IEEE Transactions on Biometrics, Behavior, and Identity Science Publication Information IEEE Transactions on Biometrics, Behavior, and Identity Science Information for Authors GaitDFG: A Deformation Field-Guided Feature Learning Framework for Gait Recognition
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1