Cargando…
Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning
With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099892/ https://www.ncbi.nlm.nih.gov/pubmed/35591124 http://dx.doi.org/10.3390/s22093433 |
_version_ | 1784706718741037056 |
---|---|
author | Choi, Won-Hyuk Choi, Yong-Suk |
author_facet | Choi, Won-Hyuk Choi, Yong-Suk |
author_sort | Choi, Won-Hyuk |
collection | PubMed |
description | With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence (CI). This is a reuse method that combines pre-trained models for different tasks. Since the CI uses a well-trained model, good performance and small training cost can be expected in the target task. We applied the CI to the Image Captioning task. Compared to using a trained feature extractor, the caption generator is usually trained from scratch. On the other hand, we pre-trained the Transformer model as a caption generator and applied CI, i.e., we used a pre-trained feature extractor and a pre-trained caption generator. To compare the training cost of the From Scratch model and the CI model, early stopping was applied during fine-tuning of the image captioning task. On the MS-COCO dataset, the vanilla image captioning model reduced training cost by 13.8% and improved performance by up to 3.2%, and the Object Relation Transformer model reduced training cost by 21.3%. |
format | Online Article Text |
id | pubmed-9099892 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-90998922022-05-14 Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning Choi, Won-Hyuk Choi, Yong-Suk Sensors (Basel) Article With the increase in the performance of deep learning models, the model parameter has increased exponentially. An increase in model parameters leads to an increase in computation and training time, i.e., an increase in training cost. To reduce the training cost, we propose Compositional Intelligence (CI). This is a reuse method that combines pre-trained models for different tasks. Since the CI uses a well-trained model, good performance and small training cost can be expected in the target task. We applied the CI to the Image Captioning task. Compared to using a trained feature extractor, the caption generator is usually trained from scratch. On the other hand, we pre-trained the Transformer model as a caption generator and applied CI, i.e., we used a pre-trained feature extractor and a pre-trained caption generator. To compare the training cost of the From Scratch model and the CI model, early stopping was applied during fine-tuning of the image captioning task. On the MS-COCO dataset, the vanilla image captioning model reduced training cost by 13.8% and improved performance by up to 3.2%, and the Object Relation Transformer model reduced training cost by 21.3%. MDPI 2022-04-30 /pmc/articles/PMC9099892/ /pubmed/35591124 http://dx.doi.org/10.3390/s22093433 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Choi, Won-Hyuk Choi, Yong-Suk Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title | Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title_full | Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title_fullStr | Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title_full_unstemmed | Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title_short | Effective Pre-Training Method and Its Compositional Intelligence for Image Captioning |
title_sort | effective pre-training method and its compositional intelligence for image captioning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099892/ https://www.ncbi.nlm.nih.gov/pubmed/35591124 http://dx.doi.org/10.3390/s22093433 |
work_keys_str_mv | AT choiwonhyuk effectivepretrainingmethodanditscompositionalintelligenceforimagecaptioning AT choiyongsuk effectivepretrainingmethodanditscompositionalintelligenceforimagecaptioning |