Cargando…

Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning

With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence...

Descripción completa

Detalles Bibliográficos
Autores principales: Choi, Won-Hyuk, Choi, Yong-Suk
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099892/
https://www.ncbi.nlm.nih.gov/pubmed/35591124
http://dx.doi.org/10.3390/s22093433
_version_ 1784706718741037056
author Choi, Won-Hyuk
Choi, Yong-Suk
author_facet Choi, Won-Hyuk
Choi, Yong-Suk
author_sort Choi, Won-Hyuk
collection PubMed
description With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence (CI). This is a reuse method that combines pre-trained models for different tasks. Since the CI uses a well-trained model, good performance and small training cost can be expected in the target task. We applied the CI to the Image Captioning task. Compared to using a trained feature extractor, the caption generator is usually trained from scratch. On the other hand, we pre-trained the Transformer model as a caption generator and applied CI, i.e., we used a pre-trained feature extractor and a pre-trained caption generator. To compare the training cost of the From Scratch model and the CI model, early stopping was applied during fine-tuning of the image captioning task. On the MS-COCO dataset, the vanilla image captioning model reduced training cost by 13.8% and improved performance by up to 3.2%, and the Object Relation Transformer model reduced training cost by 21.3%.
format Online
Article
Text
id pubmed-9099892
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-90998922022-05-14 Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning Choi, Won-Hyuk Choi, Yong-Suk Sensors (Basel) Article With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence (CI). This is a reuse method that combines pre-trained models for different tasks. Since the CI uses a well-trained model, good performance and small training cost can be expected in the target task. We applied the CI to the Image Captioning task. Compared to using a trained feature extractor, the caption generator is usually trained from scratch. On the other hand, we pre-trained the Transformer model as a caption generator and applied CI, i.e., we used a pre-trained feature extractor and a pre-trained caption generator. To compare the training cost of the From Scratch model and the CI model, early stopping was applied during fine-tuning of the image captioning task. On the MS-COCO dataset, the vanilla image captioning model reduced training cost by 13.8% and improved performance by up to 3.2%, and the Object Relation Transformer model reduced training cost by 21.3%. MDPI 2022-04-30 /pmc/articles/PMC9099892/ /pubmed/35591124 http://dx.doi.org/10.3390/s22093433 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Choi, Won-Hyuk
Choi, Yong-Suk
Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title_full Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title_fullStr Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title_full_unstemmed Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title_short Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
title_sort effective pre-training method and its compositional intelligence for image captioning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099892/
https://www.ncbi.nlm.nih.gov/pubmed/35591124
http://dx.doi.org/10.3390/s22093433
work_keys_str_mv AT choiwonhyuk effectivepretrainingmethodanditscompositionalintelligenceforimagecaptioning
AT choiyongsuk effectivepretrainingmethodanditscompositionalintelligenceforimagecaptioning