Cargando…
Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process
In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitt...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961777/ https://www.ncbi.nlm.nih.gov/pubmed/33800810 http://dx.doi.org/10.3390/s21051846 |
_version_ | 1783665335756390400 |
---|---|
author | Saffari, Mohsen Khodayar, Mahdi Ebrahimi Saadabadi, Mohammad Saeed Sequeira, Ana F. Cardoso, Jaime S. |
author_facet | Saffari, Mohsen Khodayar, Mahdi Ebrahimi Saadabadi, Mohammad Saeed Sequeira, Ana F. Cardoso, Jaime S. |
author_sort | Saffari, Mohsen |
collection | PubMed |
description | In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitting that has witnessed significant success in various classification applications. Recently, many efforts have been made to improve the Standard dropout using an unsupervised merit-based semantic selection of neurons in the latent space. However, these studies do not consider the task-relevant information quality and quantity and the diversity of the latent kernels. To solve the challenge of dropping less informative neurons in deep learning, we propose an efficient end-to-end dropout algorithm that selects the most informative neurons with the highest correlation with the target output considering the sparsity in its selection procedure. First, to promote activation diversity, we devise an approach to select the most diverse set of neurons by making use of determinantal point process (DPP) sampling. Furthermore, to incorporate task specificity into deep latent features, a mutual information (MI)-based merit function is developed. Leveraging the proposed MI with DPP sampling, we introduce the novel DPPMI dropout that adaptively adjusts the retention rate of neurons based on their contribution to the neural network task. Empirical studies on real-world classification benchmarks including, MNIST, SVHN, CIFAR10, CIFAR100, demonstrate the superiority of our proposed method over recent state-of-the-art dropout algorithms in the literature. |
format | Online Article Text |
id | pubmed-7961777 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-79617772021-03-17 Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process Saffari, Mohsen Khodayar, Mahdi Ebrahimi Saadabadi, Mohammad Saeed Sequeira, Ana F. Cardoso, Jaime S. Sensors (Basel) Article In recent years, deep neural networks have shown significant progress in computer vision due to their large generalization capacity; however, the overfitting problem ubiquitously threatens the learning process of these highly nonlinear architectures. Dropout is a recent solution to mitigate overfitting that has witnessed significant success in various classification applications. Recently, many efforts have been made to improve the Standard dropout using an unsupervised merit-based semantic selection of neurons in the latent space. However, these studies do not consider the task-relevant information quality and quantity and the diversity of the latent kernels. To solve the challenge of dropping less informative neurons in deep learning, we propose an efficient end-to-end dropout algorithm that selects the most informative neurons with the highest correlation with the target output considering the sparsity in its selection procedure. First, to promote activation diversity, we devise an approach to select the most diverse set of neurons by making use of determinantal point process (DPP) sampling. Furthermore, to incorporate task specificity into deep latent features, a mutual information (MI)-based merit function is developed. Leveraging the proposed MI with DPP sampling, we introduce the novel DPPMI dropout that adaptively adjusts the retention rate of neurons based on their contribution to the neural network task. Empirical studies on real-world classification benchmarks including, MNIST, SVHN, CIFAR10, CIFAR100, demonstrate the superiority of our proposed method over recent state-of-the-art dropout algorithms in the literature. MDPI 2021-03-06 /pmc/articles/PMC7961777/ /pubmed/33800810 http://dx.doi.org/10.3390/s21051846 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Saffari, Mohsen Khodayar, Mahdi Ebrahimi Saadabadi, Mohammad Saeed Sequeira, Ana F. Cardoso, Jaime S. Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title | Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title_full | Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title_fullStr | Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title_full_unstemmed | Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title_short | Maximum Relevance Minimum Redundancy Dropout with Informative Kernel Determinantal Point Process |
title_sort | maximum relevance minimum redundancy dropout with informative kernel determinantal point process |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961777/ https://www.ncbi.nlm.nih.gov/pubmed/33800810 http://dx.doi.org/10.3390/s21051846 |
work_keys_str_mv | AT saffarimohsen maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess AT khodayarmahdi maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess AT ebrahimisaadabadimohammadsaeed maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess AT sequeiraanaf maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess AT cardosojaimes maximumrelevanceminimumredundancydropoutwithinformativekerneldeterminantalpointprocess |