Cargando…

Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process

In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitt...

Descripción completa

Detalles Bibliográficos
Autores principales: Saffari, Mohsen, Khodayar, Mahdi, Ebrahimi Saadabadi, Mohammad Saeed, Sequeira, Ana F., Cardoso, Jaime S.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961777/
https://www.ncbi.nlm.nih.gov/pubmed/33800810
http://dx.doi.org/10.3390/s21051846
_version_ 1783665335756390400
author Saffari, Mohsen
Khodayar, Mahdi
Ebrahimi Saadabadi, Mohammad Saeed
Sequeira, Ana F.
Cardoso, Jaime S.
author_facet Saffari, Mohsen
Khodayar, Mahdi
Ebrahimi Saadabadi, Mohammad Saeed
Sequeira, Ana F.
Cardoso, Jaime S.
author_sort Saffari, Mohsen
collection PubMed
description In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitting that has witnessed significant success in various classification applications. Recently, many efforts have been made to improve the Standard dropout using an unsupervised merit-based semantic selection of neurons in the latent space. However, these studies do not consider the task-relevant information quality and quantity and the diversity of the latent kernels. To solve the challenge of dropping less informative neurons in deep learning, we propose an efficient end-to-end dropout algorithm that selects the most informative neurons with the highest correlation with the target output considering the sparsity in its selection procedure. First, to promote activation diversity, we devise an approach to select the most diverse set of neurons by making use of determinantal point process (DPP) sampling. Furthermore, to incorporate task specificity into deep latent features, a mutual information (MI)-based merit function is developed. Leveraging the proposed MI with DPP sampling, we introduce the novel DPPMI dropout that adaptively adjusts the retention rate of neurons based on their contribution to the neural network task. Empirical studies on real-world classification benchmarks including, MNIST, SVHN, CIFAR10, CIFAR100, demonstrate the superiority of our proposed method over recent state-of-the-art dropout algorithms in the literature.
format Online
Article
Text
id pubmed-7961777
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-79617772021-03-17 Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process Saffari, Mohsen Khodayar, Mahdi Ebrahimi Saadabadi, Mohammad Saeed Sequeira, Ana F. Cardoso, Jaime S. Sensors (Basel) Article In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitting that has witnessed significant success in various classification applications. Recently, many efforts have been made to improve the Standard dropout using an unsupervised merit-based semantic selection of neurons in the latent space. However, these studies do not consider the task-relevant information quality and quantity and the diversity of the latent kernels. To solve the challenge of dropping less informative neurons in deep learning, we propose an efficient end-to-end dropout algorithm that selects the most informative neurons with the highest correlation with the target output considering the sparsity in its selection procedure. First, to promote activation diversity, we devise an approach to select the most diverse set of neurons by making use of determinantal point process (DPP) sampling. Furthermore, to incorporate task specificity into deep latent features, a mutual information (MI)-based merit function is developed. Leveraging the proposed MI with DPP sampling, we introduce the novel DPPMI dropout that adaptively adjusts the retention rate of neurons based on their contribution to the neural network task. Empirical studies on real-world classification benchmarks including, MNIST, SVHN, CIFAR10, CIFAR100, demonstrate the superiority of our proposed method over recent state-of-the-art dropout algorithms in the literature. MDPI 2021-03-06 /pmc/articles/PMC7961777/ /pubmed/33800810 http://dx.doi.org/10.3390/s21051846 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Saffari, Mohsen
Khodayar, Mahdi
Ebrahimi Saadabadi, Mohammad Saeed
Sequeira, Ana F.
Cardoso, Jaime S.
Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title_full Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title_fullStr Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title_full_unstemmed Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title_short Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
title_sort maximum relevance minimum redundancy dropout with informative kernel determinantal point process
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961777/
https://www.ncbi.nlm.nih.gov/pubmed/33800810
http://dx.doi.org/10.3390/s21051846
work_keys_str_mv AT saffarimohsen maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess
AT khodayarmahdi maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess
AT ebrahimisaadabadimohammadsaeed maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess
AT sequeiraanaf maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess
AT cardosojaimes maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess