Cargando…
Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516792/ https://www.ncbi.nlm.nih.gov/pubmed/33286108 http://dx.doi.org/10.3390/e22030334 |
_version_ | 1783587083025121280 |
---|---|
author | Gajowniczek, Krzysztof Liang, Yitao Friedman, Tal Ząbkowski, Tomasz Van den Broeck, Guy |
author_facet | Gajowniczek, Krzysztof Liang, Yitao Friedman, Tal Ząbkowski, Tomasz Van den Broeck, Guy |
author_sort | Gajowniczek, Krzysztof |
collection | PubMed |
description | The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information on the distribution of the examples or, simultaneously, an extension of unsupervised learning guided by some constraints. In this article we present a methodology that bridges between artificial neural network output vectors and logical constraints. In order to do this, we present a semantic loss function and a generalized entropy loss function (Rényi entropy) that capture how close the neural network is to satisfying the constraints on its output. Our methods are intended to be generally applicable and compatible with any feedforward neural network. Therefore, the semantic loss and generalized entropy loss are simply a regularization term that can be directly plugged into an existing loss function. We evaluate our methodology over an artificially simulated dataset and two commonly used benchmark datasets which are MNIST and Fashion-MNIST to assess the relation between the analyzed loss functions and the influence of the various input and tuning parameters on the classification accuracy. The experimental evaluation shows that both losses effectively guide the learner to achieve (near-) state-of-the-art results on semi-supervised multiclass classification. |
format | Online Article Text |
id | pubmed-7516792 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75167922020-11-09 Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning Gajowniczek, Krzysztof Liang, Yitao Friedman, Tal Ząbkowski, Tomasz Van den Broeck, Guy Entropy (Basel) Article The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information on the distribution of the examples or, simultaneously, an extension of unsupervised learning guided by some constraints. In this article we present a methodology that bridges between artificial neural network output vectors and logical constraints. In order to do this, we present a semantic loss function and a generalized entropy loss function (Rényi entropy) that capture how close the neural network is to satisfying the constraints on its output. Our methods are intended to be generally applicable and compatible with any feedforward neural network. Therefore, the semantic loss and generalized entropy loss are simply a regularization term that can be directly plugged into an existing loss function. We evaluate our methodology over an artificially simulated dataset and two commonly used benchmark datasets which are MNIST and Fashion-MNIST to assess the relation between the analyzed loss functions and the influence of the various input and tuning parameters on the classification accuracy. The experimental evaluation shows that both losses effectively guide the learner to achieve (near-) state-of-the-art results on semi-supervised multiclass classification. MDPI 2020-03-14 /pmc/articles/PMC7516792/ /pubmed/33286108 http://dx.doi.org/10.3390/e22030334 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Gajowniczek, Krzysztof Liang, Yitao Friedman, Tal Ząbkowski, Tomasz Van den Broeck, Guy Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title | Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title_full | Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title_fullStr | Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title_full_unstemmed | Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title_short | Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning |
title_sort | semantic and generalized entropy loss functions for semi-supervised deep learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516792/ https://www.ncbi.nlm.nih.gov/pubmed/33286108 http://dx.doi.org/10.3390/e22030334 |
work_keys_str_mv | AT gajowniczekkrzysztof semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning AT liangyitao semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning AT friedmantal semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning AT zabkowskitomasz semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning AT vandenbroeckguy semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning |