Cargando…

Deep Realistic Facial Editing via Label-restricted Mask Disentanglement

With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute man...

Descripción completa

Detalles Bibliográficos
Autores principales: Song, Jiaming, Tong, Fenghua, Chen, Zixun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9711998/
https://www.ncbi.nlm.nih.gov/pubmed/36465954
http://dx.doi.org/10.1155/2022/5652730
_version_ 1784841700394401792
author Song, Jiaming
Tong, Fenghua
Chen, Zixun
author_facet Song, Jiaming
Tong, Fenghua
Chen, Zixun
author_sort Song, Jiaming
collection PubMed
description With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute manipulation. However, they either adopt reference-guided translation methods for manipulation or monolithic modules for diverse attribute exchange, which cannot accurately disentangle the exact facial attributes with specific styles from the reference image. In this paper, we propose a deep realistic facial editing method (termed LMGAN) based on target region focusing and dual label constraint. The proposed method, manipulating target attributes by latent space exchange, consists of subnetworks for every individual attribute. Each subnetwork exerts label-restrictions on both the target attributes exchanging stage and the training process aimed at optimizing generative quality and reference-style correlation. Our method performs greatly on disentangled representation and transferring the target attribute's style accurately. A global discriminator is introduced to combine the generated editing regional image with other nonediting areas of the source image. Both qualitative and quantitative results on the CelebA dataset verify the ability of the proposed LMGAN.
format Online
Article
Text
id pubmed-9711998
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-97119982022-12-01 Deep Realistic Facial Editing via Label-restricted Mask Disentanglement Song, Jiaming Tong, Fenghua Chen, Zixun Comput Intell Neurosci Research Article With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute manipulation. However, they either adopt reference-guided translation methods for manipulation or monolithic modules for diverse attribute exchange, which cannot accurately disentangle the exact facial attributes with specific styles from the reference image. In this paper, we propose a deep realistic facial editing method (termed LMGAN) based on target region focusing and dual label constraint. The proposed method, manipulating target attributes by latent space exchange, consists of subnetworks for every individual attribute. Each subnetwork exerts label-restrictions on both the target attributes exchanging stage and the training process aimed at optimizing generative quality and reference-style correlation. Our method performs greatly on disentangled representation and transferring the target attribute's style accurately. A global discriminator is introduced to combine the generated editing regional image with other nonediting areas of the source image. Both qualitative and quantitative results on the CelebA dataset verify the ability of the proposed LMGAN. Hindawi 2022-11-23 /pmc/articles/PMC9711998/ /pubmed/36465954 http://dx.doi.org/10.1155/2022/5652730 Text en Copyright © 2022 Jiaming Song et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Song, Jiaming
Tong, Fenghua
Chen, Zixun
Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title_full Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title_fullStr Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title_full_unstemmed Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title_short Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
title_sort deep realistic facial editing via label-restricted mask disentanglement
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9711998/
https://www.ncbi.nlm.nih.gov/pubmed/36465954
http://dx.doi.org/10.1155/2022/5652730
work_keys_str_mv AT songjiaming deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement
AT tongfenghua deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement
AT chenzixun deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement