Cargando…
An experimental study of animating-based facial image manipulation in online class environments
Recent advances in artificial intelligence technology have significantly improved facial image manipulation, which is known as Deepfake. Facial image manipulation synthesizes or replaces a region of the face in an image with that of another face. The techniques for facial image manipulation are clas...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10031167/ https://www.ncbi.nlm.nih.gov/pubmed/36949178 http://dx.doi.org/10.1038/s41598-023-31408-y |
_version_ | 1784910544670556160 |
---|---|
author | Park, Jeong-Ha Lim, Chae-Yun Kwon, Hyuk-Yoon |
author_facet | Park, Jeong-Ha Lim, Chae-Yun Kwon, Hyuk-Yoon |
author_sort | Park, Jeong-Ha |
collection | PubMed |
description | Recent advances in artificial intelligence technology have significantly improved facial image manipulation, which is known as Deepfake. Facial image manipulation synthesizes or replaces a region of the face in an image with that of another face. The techniques for facial image manipulation are classified into four categories: (1) entire face synthesis, (2) identity swap, (3) attribute manipulation, and (4) expression swap. Out of them, we focus on expression swap because it effectively manipulates only the expression of the face in the images or videos without creating or replacing the entire face, having advantages for the real-time application. In this study, we propose an evaluation framework of the expression swap models targeting the real-time online class environments. For this, we define three kinds of scenarios according to the portion of the face in the entire image considering actual online class situations: (1) attendance check (Scenario 1), (2) presentation (Scenario 2), and (3) examination (Scenario 3). Considering the manipulation on the online class environments, the framework receives a single source image and a target video and generates the video that manipulates a face of the target video to that in the source image. To this end, we select two models that satisfy the conditions required by the framework: (1) first order model and (2) GANimation. We implement these models in the framework and evaluate their performance for the defined scenarios. Through the quantitative and qualitative evaluation, we observe distinguishing properties of the used two models. Specifically, both models show acceptable results in Scenario 1, where the face occupies a large portion of the image. However, their performances are significantly degraded in Scenarios 2 and 3, where the face occupies less portion of the image; the first order model causes relatively less loss of image quality than GANimation in the result of the quantitative evaluation. In contrast, GANimation has the advantages of representing facial expression changes compared to the first order model. Finally, we devise an architecture for applying the expression swap model to the online video conferencing application in real-time. In particular, by applying the expression swap model to widely used online meeting platforms such as Zoom, Google Meet, and Microsoft Teams, we demonstrate its feasibility for real-time online classes. |
format | Online Article Text |
id | pubmed-10031167 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-100311672023-03-22 An experimental study of animating-based facial image manipulation in online class environments Park, Jeong-Ha Lim, Chae-Yun Kwon, Hyuk-Yoon Sci Rep Article Recent advances in artificial intelligence technology have significantly improved facial image manipulation, which is known as Deepfake. Facial image manipulation synthesizes or replaces a region of the face in an image with that of another face. The techniques for facial image manipulation are classified into four categories: (1) entire face synthesis, (2) identity swap, (3) attribute manipulation, and (4) expression swap. Out of them, we focus on expression swap because it effectively manipulates only the expression of the face in the images or videos without creating or replacing the entire face, having advantages for the real-time application. In this study, we propose an evaluation framework of the expression swap models targeting the real-time online class environments. For this, we define three kinds of scenarios according to the portion of the face in the entire image considering actual online class situations: (1) attendance check (Scenario 1), (2) presentation (Scenario 2), and (3) examination (Scenario 3). Considering the manipulation on the online class environments, the framework receives a single source image and a target video and generates the video that manipulates a face of the target video to that in the source image. To this end, we select two models that satisfy the conditions required by the framework: (1) first order model and (2) GANimation. We implement these models in the framework and evaluate their performance for the defined scenarios. Through the quantitative and qualitative evaluation, we observe distinguishing properties of the used two models. Specifically, both models show acceptable results in Scenario 1, where the face occupies a large portion of the image. However, their performances are significantly degraded in Scenarios 2 and 3, where the face occupies less portion of the image; the first order model causes relatively less loss of image quality than GANimation in the result of the quantitative evaluation. In contrast, GANimation has the advantages of representing facial expression changes compared to the first order model. Finally, we devise an architecture for applying the expression swap model to the online video conferencing application in real-time. In particular, by applying the expression swap model to widely used online meeting platforms such as Zoom, Google Meet, and Microsoft Teams, we demonstrate its feasibility for real-time online classes. Nature Publishing Group UK 2023-03-22 /pmc/articles/PMC10031167/ /pubmed/36949178 http://dx.doi.org/10.1038/s41598-023-31408-y Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Park, Jeong-Ha Lim, Chae-Yun Kwon, Hyuk-Yoon An experimental study of animating-based facial image manipulation in online class environments |
title | An experimental study of animating-based facial image manipulation in online class environments |
title_full | An experimental study of animating-based facial image manipulation in online class environments |
title_fullStr | An experimental study of animating-based facial image manipulation in online class environments |
title_full_unstemmed | An experimental study of animating-based facial image manipulation in online class environments |
title_short | An experimental study of animating-based facial image manipulation in online class environments |
title_sort | experimental study of animating-based facial image manipulation in online class environments |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10031167/ https://www.ncbi.nlm.nih.gov/pubmed/36949178 http://dx.doi.org/10.1038/s41598-023-31408-y |
work_keys_str_mv | AT parkjeongha anexperimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments AT limchaeyun anexperimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments AT kwonhyukyoon anexperimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments AT parkjeongha experimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments AT limchaeyun experimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments AT kwonhyukyoon experimentalstudyofanimatingbasedfacialimagemanipulationinonlineclassenvironments |