Cargando…

Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content

Applying machine learning (ML), and especially deep learning, to understand visual content is becoming common practice in many application areas. However, little attention has been given to its use within the multimedia creative domain. It is true that ML is already popular for content creation, but...

Descripción completa

Detalles Bibliográficos
Autores principales: Viana, Paula, Andrade, Maria Teresa, Carvalho, Pedro, Vilaça, Luis, Teixeira, Inês N., Costa, Tiago, Jonker, Pieter
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8953642/
https://www.ncbi.nlm.nih.gov/pubmed/35324623
http://dx.doi.org/10.3390/jimaging8030068
_version_ 1784675901704765440
author Viana, Paula
Andrade, Maria Teresa
Carvalho, Pedro
Vilaça, Luis
Teixeira, Inês N.
Costa, Tiago
Jonker, Pieter
author_facet Viana, Paula
Andrade, Maria Teresa
Carvalho, Pedro
Vilaça, Luis
Teixeira, Inês N.
Costa, Tiago
Jonker, Pieter
author_sort Viana, Paula
collection PubMed
description Applying machine learning (ML), and especially deep learning, to understand visual content is becoming common practice in many application areas. However, little attention has been given to its use within the multimedia creative domain. It is true that ML is already popular for content creation, but the progress achieved so far addresses essentially textual content or the identification and selection of specific types of content. A wealth of possibilities are yet to be explored by bringing the use of ML into the multimedia creative process, allowing the knowledge inferred by the former to influence automatically how new multimedia content is created. The work presented in this article provides contributions in three distinct ways towards this goal: firstly, it proposes a methodology to re-train popular neural network models in identifying new thematic concepts in static visual content and attaching meaningful annotations to the detected regions of interest; secondly, it presents varied visual digital effects and corresponding tools that can be automatically called upon to apply such effects in a previously analyzed photo; thirdly, it defines a complete automated creative workflow, from the acquisition of a photograph and corresponding contextual data, through the ML region-based annotation, to the automatic application of digital effects and generation of a semantically aware multimedia story driven by the previously derived situational and visual contextual data. Additionally, it presents a variant of this automated workflow by offering to the user the possibility of manipulating the automatic annotations in an assisted manner. The final aim is to transform a static digital photo into a short video clip, taking into account the information acquired. The final result strongly contrasts with current standard approaches of creating random movements, by implementing an intelligent content- and context-aware video.
format Online
Article
Text
id pubmed-8953642
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89536422022-03-26 Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content Viana, Paula Andrade, Maria Teresa Carvalho, Pedro Vilaça, Luis Teixeira, Inês N. Costa, Tiago Jonker, Pieter J Imaging Article Applying machine learning (ML), and especially deep learning, to understand visual content is becoming common practice in many application areas. However, little attention has been given to its use within the multimedia creative domain. It is true that ML is already popular for content creation, but the progress achieved so far addresses essentially textual content or the identification and selection of specific types of content. A wealth of possibilities are yet to be explored by bringing the use of ML into the multimedia creative process, allowing the knowledge inferred by the former to influence automatically how new multimedia content is created. The work presented in this article provides contributions in three distinct ways towards this goal: firstly, it proposes a methodology to re-train popular neural network models in identifying new thematic concepts in static visual content and attaching meaningful annotations to the detected regions of interest; secondly, it presents varied visual digital effects and corresponding tools that can be automatically called upon to apply such effects in a previously analyzed photo; thirdly, it defines a complete automated creative workflow, from the acquisition of a photograph and corresponding contextual data, through the ML region-based annotation, to the automatic application of digital effects and generation of a semantically aware multimedia story driven by the previously derived situational and visual contextual data. Additionally, it presents a variant of this automated workflow by offering to the user the possibility of manipulating the automatic annotations in an assisted manner. The final aim is to transform a static digital photo into a short video clip, taking into account the information acquired. The final result strongly contrasts with current standard approaches of creating random movements, by implementing an intelligent content- and context-aware video. MDPI 2022-03-10 /pmc/articles/PMC8953642/ /pubmed/35324623 http://dx.doi.org/10.3390/jimaging8030068 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Viana, Paula
Andrade, Maria Teresa
Carvalho, Pedro
Vilaça, Luis
Teixeira, Inês N.
Costa, Tiago
Jonker, Pieter
Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title_full Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title_fullStr Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title_full_unstemmed Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title_short Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content
title_sort photo2video: semantic-aware deep learning-based video generation from still content
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8953642/
https://www.ncbi.nlm.nih.gov/pubmed/35324623
http://dx.doi.org/10.3390/jimaging8030068
work_keys_str_mv AT vianapaula photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT andrademariateresa photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT carvalhopedro photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT vilacaluis photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT teixeirainesn photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT costatiago photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent
AT jonkerpieter photo2videosemanticawaredeeplearningbasedvideogenerationfromstillcontent