Cargando…
Deep Realistic Facial Editing via Label-restricted Mask Disentanglement
With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute man...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9711998/ https://www.ncbi.nlm.nih.gov/pubmed/36465954 http://dx.doi.org/10.1155/2022/5652730 |
_version_ | 1784841700394401792 |
---|---|
author | Song, Jiaming Tong, Fenghua Chen, Zixun |
author_facet | Song, Jiaming Tong, Fenghua Chen, Zixun |
author_sort | Song, Jiaming |
collection | PubMed |
description | With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute manipulation. However, they either adopt reference-guided translation methods for manipulation or monolithic modules for diverse attribute exchange, which cannot accurately disentangle the exact facial attributes with specific styles from the reference image. In this paper, we propose a deep realistic facial editing method (termed LMGAN) based on target region focusing and dual label constraint. The proposed method, manipulating target attributes by latent space exchange, consists of subnetworks for every individual attribute. Each subnetwork exerts label-restrictions on both the target attributes exchanging stage and the training process aimed at optimizing generative quality and reference-style correlation. Our method performs greatly on disentangled representation and transferring the target attribute's style accurately. A global discriminator is introduced to combine the generated editing regional image with other nonediting areas of the source image. Both qualitative and quantitative results on the CelebA dataset verify the ability of the proposed LMGAN. |
format | Online Article Text |
id | pubmed-9711998 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-97119982022-12-01 Deep Realistic Facial Editing via Label-restricted Mask Disentanglement Song, Jiaming Tong, Fenghua Chen, Zixun Comput Intell Neurosci Research Article With the rapid development of GAN (generative adversarial network), recent years have witnessed an increasing number of tasks on reference-guided facial attributes transfer. Most state-of-the-art methods consist of facial information extraction, latent space disentanglement, and target attribute manipulation. However, they either adopt reference-guided translation methods for manipulation or monolithic modules for diverse attribute exchange, which cannot accurately disentangle the exact facial attributes with specific styles from the reference image. In this paper, we propose a deep realistic facial editing method (termed LMGAN) based on target region focusing and dual label constraint. The proposed method, manipulating target attributes by latent space exchange, consists of subnetworks for every individual attribute. Each subnetwork exerts label-restrictions on both the target attributes exchanging stage and the training process aimed at optimizing generative quality and reference-style correlation. Our method performs greatly on disentangled representation and transferring the target attribute's style accurately. A global discriminator is introduced to combine the generated editing regional image with other nonediting areas of the source image. Both qualitative and quantitative results on the CelebA dataset verify the ability of the proposed LMGAN. Hindawi 2022-11-23 /pmc/articles/PMC9711998/ /pubmed/36465954 http://dx.doi.org/10.1155/2022/5652730 Text en Copyright © 2022 Jiaming Song et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Song, Jiaming Tong, Fenghua Chen, Zixun Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title | Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title_full | Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title_fullStr | Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title_full_unstemmed | Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title_short | Deep Realistic Facial Editing via Label-restricted Mask Disentanglement |
title_sort | deep realistic facial editing via label-restricted mask disentanglement |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9711998/ https://www.ncbi.nlm.nih.gov/pubmed/36465954 http://dx.doi.org/10.1155/2022/5652730 |
work_keys_str_mv | AT songjiaming deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement AT tongfenghua deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement AT chenzixun deeprealisticfacialeditingvialabelrestrictedmaskdisentanglement |