首页 | 本学科首页   官方微博 | 高级检索  
     

基于 Swin Transformer 和 Style-based Generator 的盲人脸修复
引用本文:向泽林,楼旭东,李旭伟. 基于 Swin Transformer 和 Style-based Generator 的盲人脸修复[J]. 四川大学学报(自然科学版), 2023, 60(3): 032003
作者姓名:向泽林  楼旭东  李旭伟
作者单位:四川外国语大学成都学院,四川大学计算机学院,四川大学计算机学院
基金项目:国家重点研发项目(2020YFC0832404)
摘    要:盲人脸修复任务是从低质量的图像(例如模糊、噪声和压缩图像)中恢复高质量的图像.由于事先不知道低质量图像的退化类型和退化参数,因此盲人脸修复是一个高度病态的问题,在修复过程中严重依赖各种先验指导.然而,由于面部成分和面部标志等面部先验通常是从低质量图像中提取或估计的,可能存在不准确的情况,这直接影响最终的修复性能,因此难以有效利用这些先验知识.此外,目前的主流方法基本都是依赖ConvNets进行特征提取,没有很好地考虑长距离特征,导致最终结果缺乏连续一致性.本文提出了一种改进的StyleGAN模型,命名为SwinStyleGAN,应用在高级视觉任务上表现出色的Swin Transformer来提取长距离特征,并通过改进后的类StyleGAN合成网络逐步生成图像.本文设计了一个空间注意力转换模块SAT来重新分配每个阶段特征的像素权重,以进一步约束生成器.大量实验表明,本文提出的方法具有更好的盲人脸修复性能.

关 键 词:盲人脸修复;ConvNets;Swin Transformer;StyleGAN;空间注意力转换模块
收稿时间:2023-02-16
修稿时间:2023-03-19

Blind face restoration based on Swin Transformer and Style-Based Generator
XIANG Ze-Lin,LOU Xu-Dong and LI Xu-Wei. Blind face restoration based on Swin Transformer and Style-Based Generator[J]. Journal of Sichuan University (Natural Science Edition), 2023, 60(3): 032003
Authors:XIANG Ze-Lin  LOU Xu-Dong  LI Xu-Wei
Affiliation:Chengdu Institute Sichuan International Studies University,College of Computer Science, Sichuan University,College of Computer Science, Sichuan University
Abstract:Blind face restoration is the process of restoring a high-quality image from a low-quality image (e.g., blurred, noisy, or compressed image). Since the degradation type and degradation parameters of the low-quality image are unknown, blind face restoration is a highly ill-posed problem that heavily relies on various facial prior such as facial components and facial landmarks during the restoration process. However, these facial priors are typically extracted or estimated from low-quality images, which may be inaccurate, directly affecting the final restoration performance. The current mainstream methods mostly use ConNets for feature extraction and do not consider long-distance features, resulting in a lack of continuous consistency in the final results.The authors propose an improved StyleGAN model named SwinStyleGAN, which uses Swin Transformer to extract long-distance features and gradually generates images through an improved StyleGAN synthesis network.Addtionally, the authors design a Spatial Attention Transformation (SAT) module to reassign pixel weights of each stage feature to further constrain the generator. Experiments show that the proposed SwinStyleGAN in this paper has better blind face restoration performance.
Keywords:Blind face restoration   ConvNets   Swin Transformer   StyleGAN   Spatial attention transformation
点击此处可从《四川大学学报(自然科学版)》浏览原始摘要信息
点击此处可从《四川大学学报(自然科学版)》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号