Cargando…

Single Image Video Prediction with Auto-Regressive GANs

In this paper, we introduce an approach for future frames prediction based on a single input image. Our method is able to generate an entire video sequence based on the information contained in the input frame. We adopt an autoregressive approach in our generation process, i.e., the output from each...

Descripción completa

Detalles Bibliográficos
Autores principales: Huang, Jiahui, Chia, Yew Ken, Yu, Samson, Yee, Kevin, Küster, Dennis, Krumhuber, Eva G., Herremans, Dorien, Roig, Gemma
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099507/
https://www.ncbi.nlm.nih.gov/pubmed/35591224
http://dx.doi.org/10.3390/s22093533
_version_ 1784706622825693184
author Huang, Jiahui
Chia, Yew Ken
Yu, Samson
Yee, Kevin
Küster, Dennis
Krumhuber, Eva G.
Herremans, Dorien
Roig, Gemma
author_facet Huang, Jiahui
Chia, Yew Ken
Yu, Samson
Yee, Kevin
Küster, Dennis
Krumhuber, Eva G.
Herremans, Dorien
Roig, Gemma
author_sort Huang, Jiahui
collection PubMed
description In this paper, we introduce an approach for future frames prediction based on a single input image. Our method is able to generate an entire video sequence based on the information contained in the input frame. We adopt an autoregressive approach in our generation process, i.e., the output from each time step is fed as the input to the next step. Unlike other video prediction methods that use “one shot” generation, our method is able to preserve much more details from the input image, while also capturing the critical pixel-level changes between the frames. We overcome the problem of generation quality degradation by introducing a “complementary mask” module in our architecture, and we show that this allows the model to only focus on the generation of the pixels that need to be changed, and to reuse those that should remain static from its previous frame. We empirically validate our methods against various video prediction models on the UT Dallas Dataset, and show that our approach is able to generate high quality realistic video sequences from one static input image. In addition, we also validate the robustness of our method by testing a pre-trained model on the unseen ADFES facial expression dataset. We also provide qualitative results of our model tested on a human action dataset: The Weizmann Action database.
format Online
Article
Text
id pubmed-9099507
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-90995072022-05-14 Single Image Video Prediction with Auto-Regressive GANs Huang, Jiahui Chia, Yew Ken Yu, Samson Yee, Kevin Küster, Dennis Krumhuber, Eva G. Herremans, Dorien Roig, Gemma Sensors (Basel) Article In this paper, we introduce an approach for future frames prediction based on a single input image. Our method is able to generate an entire video sequence based on the information contained in the input frame. We adopt an autoregressive approach in our generation process, i.e., the output from each time step is fed as the input to the next step. Unlike other video prediction methods that use “one shot” generation, our method is able to preserve much more details from the input image, while also capturing the critical pixel-level changes between the frames. We overcome the problem of generation quality degradation by introducing a “complementary mask” module in our architecture, and we show that this allows the model to only focus on the generation of the pixels that need to be changed, and to reuse those that should remain static from its previous frame. We empirically validate our methods against various video prediction models on the UT Dallas Dataset, and show that our approach is able to generate high quality realistic video sequences from one static input image. In addition, we also validate the robustness of our method by testing a pre-trained model on the unseen ADFES facial expression dataset. We also provide qualitative results of our model tested on a human action dataset: The Weizmann Action database. MDPI 2022-05-06 /pmc/articles/PMC9099507/ /pubmed/35591224 http://dx.doi.org/10.3390/s22093533 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Huang, Jiahui
Chia, Yew Ken
Yu, Samson
Yee, Kevin
Küster, Dennis
Krumhuber, Eva G.
Herremans, Dorien
Roig, Gemma
Single Image Video Prediction with Auto-Regressive GANs
title Single Image Video Prediction with Auto-Regressive GANs
title_full Single Image Video Prediction with Auto-Regressive GANs
title_fullStr Single Image Video Prediction with Auto-Regressive GANs
title_full_unstemmed Single Image Video Prediction with Auto-Regressive GANs
title_short Single Image Video Prediction with Auto-Regressive GANs
title_sort single image video prediction with auto-regressive gans
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099507/
https://www.ncbi.nlm.nih.gov/pubmed/35591224
http://dx.doi.org/10.3390/s22093533
work_keys_str_mv AT huangjiahui singleimagevideopredictionwithautoregressivegans
AT chiayewken singleimagevideopredictionwithautoregressivegans
AT yusamson singleimagevideopredictionwithautoregressivegans
AT yeekevin singleimagevideopredictionwithautoregressivegans
AT kusterdennis singleimagevideopredictionwithautoregressivegans
AT krumhuberevag singleimagevideopredictionwithautoregressivegans
AT herremansdorien singleimagevideopredictionwithautoregressivegans
AT roiggemma singleimagevideopredictionwithautoregressivegans