Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/105455
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.creatorYang, Ten_US
dc.creatorRen, Pen_US
dc.creatorXie, Xen_US
dc.creatorZhang, Len_US
dc.date.accessioned2024-04-15T07:34:29Z-
dc.date.available2024-04-15T07:34:29Z-
dc.identifier.isbn978-1-6654-4509-2 (Electronic)en_US
dc.identifier.isbn978-1-6654-4510-8 (Print on Demand(PoD))en_US
dc.identifier.urihttp://hdl.handle.net/10397/105455-
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineersen_US
dc.rights©2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.rightsThe following publication T. Yang, P. Ren, X. Xie and L. Zhang, "GAN Prior Embedded Network for Blind Face Restoration in the Wild," 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 2021, pp. 672-681 is available at https://doi.org/10.1109/CVPR46437.2021.00073.en_US
dc.titleGAN prior embedded network for blind face restoration in the wilden_US
dc.typeConference Paperen_US
dc.identifier.spage672en_US
dc.identifier.epage681en_US
dc.identifier.doi10.1109/CVPR46437.2021.00073en_US
dcterms.abstractBlind face restoration (BFR) from severely degraded face images in the wild is a very challenging problem. Due to the high illness of the problem and the complex unknown degradation, directly training a deep neural network (DNN) usually cannot lead to acceptable results. Existing generative adversarial network (GAN) based methods can produce better results but tend to generate over-smoothed restorations. In this work, we propose a new method by first learning a GAN for high-quality face image generation and embedding it into a U-shaped DNN as a prior decoder, then fine-tuning the GAN prior embedded DNN with a set of synthesized low-quality face images. The GAN blocks are designed to ensure that the latent code and noise input to the GAN can be respectively generated from the deep and shallow features of the DNN, controlling the global face structure, local face details and background of the reconstructed image. The proposed GAN prior embedded network (GPEN) is easy-to-implement, and it can generate visually photo-realistic results. Our experiments demonstrated that the proposed GPEN achieves significantly superior results to state-of-the-art BFR methods both quantitatively and qualitatively, especially for the restoration of severely degraded face images in the wild. The source code and models can be found at https://github.com/yangxy/GPEN.-
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitation2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19-25 June 2021, p. 672-681en_US
dcterms.issued2021-
dc.identifier.scopus2-s2.0-85122174696-
dc.relation.conferenceIEEE/CVF Conference on Computer Vision and Pattern Recognition [CVPR]-
dc.description.validate202402 bcch-
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumberCOMP-0037-
dc.description.fundingSourceRGCen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS56309576-
dc.description.oaCategoryGreen (AAM)en_US
Appears in Collections:Conference Paper
Files in This Item:
File Description SizeFormat 
Zhang_Gan_Prior_Embedded.pdfPre-Published version6.23 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

122
Last Week
3
Last month
Citations as of Nov 9, 2025

Downloads

130
Citations as of Nov 9, 2025

SCOPUSTM   
Citations

311
Citations as of Dec 19, 2025

WEB OF SCIENCETM
Citations

228
Citations as of Dec 18, 2025

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.