Cargando…

Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm

Some neural network can be trained by transfer learning, which uses a pre-trained neural network as the source task, for a small target task’s dataset. The performance of the transfer learning depends on the knowledge (i.e., layers) selected from the pre-trained network. At present, this knowledge i...

Descripción completa

Detalles Bibliográficos
Autores principales: Imai, Shunsuke, Kawai, Shin, Nobuhara, Hajime
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7235242/
https://www.ncbi.nlm.nih.gov/pubmed/32424180
http://dx.doi.org/10.1038/s41598-020-64165-3
_version_ 1783535923294633984
author Imai, Shunsuke
Kawai, Shin
Nobuhara, Hajime
author_facet Imai, Shunsuke
Kawai, Shin
Nobuhara, Hajime
author_sort Imai, Shunsuke
collection PubMed
description Some neural network can be trained by transfer learning, which uses a pre-trained neural network as the source task, for a small target task’s dataset. The performance of the transfer learning depends on the knowledge (i.e., layers) selected from the pre-trained network. At present, this knowledge is usually chosen by humans. The transfer learning method PathNet automatically selects pre-trained modules or adjustable modules in a modular neural network. However, PathNet requires modular neural networks as the pre-trained networks, therefore non-modular pre-trained neural networks are currently unavailable. Consequently, PathNet limits the versatility of the network structure. To address this limitation, we propose Stepwise PathNet, which regards the layers of a non-modular pre-trained neural network as the module in PathNet and selects the layers automatically through training. In an experimental validation of transfer learning from InceptionV3 pre-trained on the ImageNet dataset to networks trained on three other datasets (CIFAR-100, SVHN and Food-101), Stepwise PathNet was up to 8% and 10% more accurate than finely tuned and from-scratch approaches, respectively. Also, some of the selected layers were not supported by the layer functions assumed in PathNet.
format Online
Article
Text
id pubmed-7235242
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-72352422020-05-29 Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm Imai, Shunsuke Kawai, Shin Nobuhara, Hajime Sci Rep Article Some neural network can be trained by transfer learning, which uses a pre-trained neural network as the source task, for a small target task’s dataset. The performance of the transfer learning depends on the knowledge (i.e., layers) selected from the pre-trained network. At present, this knowledge is usually chosen by humans. The transfer learning method PathNet automatically selects pre-trained modules or adjustable modules in a modular neural network. However, PathNet requires modular neural networks as the pre-trained networks, therefore non-modular pre-trained neural networks are currently unavailable. Consequently, PathNet limits the versatility of the network structure. To address this limitation, we propose Stepwise PathNet, which regards the layers of a non-modular pre-trained neural network as the module in PathNet and selects the layers automatically through training. In an experimental validation of transfer learning from InceptionV3 pre-trained on the ImageNet dataset to networks trained on three other datasets (CIFAR-100, SVHN and Food-101), Stepwise PathNet was up to 8% and 10% more accurate than finely tuned and from-scratch approaches, respectively. Also, some of the selected layers were not supported by the layer functions assumed in PathNet. Nature Publishing Group UK 2020-05-18 /pmc/articles/PMC7235242/ /pubmed/32424180 http://dx.doi.org/10.1038/s41598-020-64165-3 Text en © The Author(s) 2020 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Imai, Shunsuke
Kawai, Shin
Nobuhara, Hajime
Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title_full Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title_fullStr Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title_full_unstemmed Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title_short Stepwise PathNet: a layer-by-layer knowledge-selection-based transfer learning algorithm
title_sort stepwise pathnet: a layer-by-layer knowledge-selection-based transfer learning algorithm
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7235242/
https://www.ncbi.nlm.nih.gov/pubmed/32424180
http://dx.doi.org/10.1038/s41598-020-64165-3
work_keys_str_mv AT imaishunsuke stepwisepathnetalayerbylayerknowledgeselectionbasedtransferlearningalgorithm
AT kawaishin stepwisepathnetalayerbylayerknowledgeselectionbasedtransferlearningalgorithm
AT nobuharahajime stepwisepathnetalayerbylayerknowledgeselectionbasedtransferlearningalgorithm