Cargando…

Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning

The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information...

Descripción completa

Detalles Bibliográficos
Autores principales: Gajowniczek, Krzysztof, Liang, Yitao, Friedman, Tal, Ząbkowski, Tomasz, Van den Broeck, Guy
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516792/
https://www.ncbi.nlm.nih.gov/pubmed/33286108
http://dx.doi.org/10.3390/e22030334
_version_ 1783587083025121280
author Gajowniczek, Krzysztof
Liang, Yitao
Friedman, Tal
Ząbkowski, Tomasz
Van den Broeck, Guy
author_facet Gajowniczek, Krzysztof
Liang, Yitao
Friedman, Tal
Ząbkowski, Tomasz
Van den Broeck, Guy
author_sort Gajowniczek, Krzysztof
collection PubMed
description The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information on the distribution of the examples or, simultaneously, an extension of unsupervised learning guided by some constraints. In this article we present a methodology that bridges between artificial neural network output vectors and logical constraints. In order to do this, we present a semantic loss function and a generalized entropy loss function (Rényi entropy) that capture how close the neural network is to satisfying the constraints on its output. Our methods are intended to be generally applicable and compatible with any feedforward neural network. Therefore, the semantic loss and generalized entropy loss are simply a regularization term that can be directly plugged into an existing loss function. We evaluate our methodology over an artificially simulated dataset and two commonly used benchmark datasets which are MNIST and Fashion-MNIST to assess the relation between the analyzed loss functions and the influence of the various input and tuning parameters on the classification accuracy. The experimental evaluation shows that both losses effectively guide the learner to achieve (near-) state-of-the-art results on semi-supervised multiclass classification.
format Online
Article
Text
id pubmed-7516792
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75167922020-11-09 Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning Gajowniczek, Krzysztof Liang, Yitao Friedman, Tal Ząbkowski, Tomasz Van den Broeck, Guy Entropy (Basel) Article The increasing size of modern datasets combined with the difficulty of obtaining real label information (e.g., class) has made semi-supervised learning a problem of considerable practical importance in modern data analysis. Semi-supervised learning is supervised learning with additional information on the distribution of the examples or, simultaneously, an extension of unsupervised learning guided by some constraints. In this article we present a methodology that bridges between artificial neural network output vectors and logical constraints. In order to do this, we present a semantic loss function and a generalized entropy loss function (Rényi entropy) that capture how close the neural network is to satisfying the constraints on its output. Our methods are intended to be generally applicable and compatible with any feedforward neural network. Therefore, the semantic loss and generalized entropy loss are simply a regularization term that can be directly plugged into an existing loss function. We evaluate our methodology over an artificially simulated dataset and two commonly used benchmark datasets which are MNIST and Fashion-MNIST to assess the relation between the analyzed loss functions and the influence of the various input and tuning parameters on the classification accuracy. The experimental evaluation shows that both losses effectively guide the learner to achieve (near-) state-of-the-art results on semi-supervised multiclass classification. MDPI 2020-03-14 /pmc/articles/PMC7516792/ /pubmed/33286108 http://dx.doi.org/10.3390/e22030334 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gajowniczek, Krzysztof
Liang, Yitao
Friedman, Tal
Ząbkowski, Tomasz
Van den Broeck, Guy
Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title_full Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title_fullStr Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title_full_unstemmed Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title_short Semantic and Generalized Entropy Loss Functions for Semi-Supervised Deep Learning
title_sort semantic and generalized entropy loss functions for semi-supervised deep learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516792/
https://www.ncbi.nlm.nih.gov/pubmed/33286108
http://dx.doi.org/10.3390/e22030334
work_keys_str_mv AT gajowniczekkrzysztof semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning
AT liangyitao semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning
AT friedmantal semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning
AT zabkowskitomasz semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning
AT vandenbroeckguy semanticandgeneralizedentropylossfunctionsforsemisuperviseddeeplearning