Cargando…

Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer

Generating images of artistic style from input images, also known as image style transfer, has been improved in the quality of output style and the speed of image generation since deep neural networks have been applied in the field of computer vision research. However, the previous approaches used f...

Descripción completa

Detalles Bibliográficos
Autor principal: Choi, Hyun-Chul
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9003536/
https://www.ncbi.nlm.nih.gov/pubmed/35408228
http://dx.doi.org/10.3390/s22072611
_version_ 1784686153176186880
author Choi, Hyun-Chul
author_facet Choi, Hyun-Chul
author_sort Choi, Hyun-Chul
collection PubMed
description Generating images of artistic style from input images, also known as image style transfer, has been improved in the quality of output style and the speed of image generation since deep neural networks have been applied in the field of computer vision research. However, the previous approaches used feature alignment techniques that were too simple in their transform layer to cover the characteristics of style features of images. In addition, they used an inconsistent combination of transform layers and loss functions in the training phase to embed arbitrary styles in a decoder network. To overcome these shortcomings, the second-order statistics of the encoded features are exploited to build an optimal arbitrary image style transfer technique. First, a new correlation-aware loss and a correlation-aware feature alignment technique are proposed. Using this consistent combination of loss and feature alignment methods strongly matches the second-order statistics of content features to those of the target-style features and, accordingly, the style capacity of the decoder network is increased. Secondly, a new component-wise style controlling method is proposed. This method can generate various styles from one or several style images by using style-specific components from second-order feature statistics. We experimentally prove that the proposed method achieves improvements in both the style capacity of the decoder network and the style variety without losing the ability of real-time processing (less than 200 ms) on Graphics Processing Unit (GPU) devices.
format Online
Article
Text
id pubmed-9003536
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-90035362022-04-13 Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer Choi, Hyun-Chul Sensors (Basel) Article Generating images of artistic style from input images, also known as image style transfer, has been improved in the quality of output style and the speed of image generation since deep neural networks have been applied in the field of computer vision research. However, the previous approaches used feature alignment techniques that were too simple in their transform layer to cover the characteristics of style features of images. In addition, they used an inconsistent combination of transform layers and loss functions in the training phase to embed arbitrary styles in a decoder network. To overcome these shortcomings, the second-order statistics of the encoded features are exploited to build an optimal arbitrary image style transfer technique. First, a new correlation-aware loss and a correlation-aware feature alignment technique are proposed. Using this consistent combination of loss and feature alignment methods strongly matches the second-order statistics of content features to those of the target-style features and, accordingly, the style capacity of the decoder network is increased. Secondly, a new component-wise style controlling method is proposed. This method can generate various styles from one or several style images by using style-specific components from second-order feature statistics. We experimentally prove that the proposed method achieves improvements in both the style capacity of the decoder network and the style variety without losing the ability of real-time processing (less than 200 ms) on Graphics Processing Unit (GPU) devices. MDPI 2022-03-29 /pmc/articles/PMC9003536/ /pubmed/35408228 http://dx.doi.org/10.3390/s22072611 Text en © 2022 by the author. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Choi, Hyun-Chul
Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title_full Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title_fullStr Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title_full_unstemmed Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title_short Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer
title_sort toward exploiting second-order feature statistics for arbitrary image style transfer
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9003536/
https://www.ncbi.nlm.nih.gov/pubmed/35408228
http://dx.doi.org/10.3390/s22072611
work_keys_str_mv AT choihyunchul towardexploitingsecondorderfeaturestatisticsforarbitraryimagestyletransfer