Cargando…

Deep forest

Current deep-learning models are mostly built upon neural networks, i.e. multiple layers of parameterized differentiable non-linear modules that can be trained by backpropagation. In this paper, we explore the possibility of building deep models based on non-differentiable modules such as decision t...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhou, Zhi-Hua, Feng, Ji
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8291612/
https://www.ncbi.nlm.nih.gov/pubmed/34691833
http://dx.doi.org/10.1093/nsr/nwy108
_version_ 1783724673517748224
author Zhou, Zhi-Hua
Feng, Ji
author_facet Zhou, Zhi-Hua
Feng, Ji
author_sort Zhou, Zhi-Hua
collection PubMed
description Current deep-learning models are mostly built upon neural networks, i.e. multiple layers of parameterized differentiable non-linear modules that can be trained by backpropagation. In this paper, we explore the possibility of building deep models based on non-differentiable modules such as decision trees. After a discussion about the mystery behind deep neural networks, particularly by contrasting them with shallow neural networks and traditional machine-learning techniques such as decision trees and boosting machines, we conjecture that the success of deep neural networks owes much to three characteristics, i.e. layer-by-layer processing, in-model feature transformation and sufficient model complexity. On one hand, our conjecture may offer inspiration for theoretical understanding of deep learning; on the other hand, to verify the conjecture, we propose an approach that generates deep forest holding these characteristics. This is a decision-tree ensemble approach, with fewer hyper-parameters than deep neural networks, and its model complexity can be automatically determined in a data-dependent way. Experiments show that its performance is quite robust to hyper-parameter settings, such that in most cases, even across different data from different domains, it is able to achieve excellent performance by using the same default setting. This study opens the door to deep learning based on non-differentiable modules without gradient-based adjustment, and exhibits the possibility of constructing deep models without backpropagation.
format Online
Article
Text
id pubmed-8291612
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-82916122021-10-21 Deep forest Zhou, Zhi-Hua Feng, Ji Natl Sci Rev Information Science Current deep-learning models are mostly built upon neural networks, i.e. multiple layers of parameterized differentiable non-linear modules that can be trained by backpropagation. In this paper, we explore the possibility of building deep models based on non-differentiable modules such as decision trees. After a discussion about the mystery behind deep neural networks, particularly by contrasting them with shallow neural networks and traditional machine-learning techniques such as decision trees and boosting machines, we conjecture that the success of deep neural networks owes much to three characteristics, i.e. layer-by-layer processing, in-model feature transformation and sufficient model complexity. On one hand, our conjecture may offer inspiration for theoretical understanding of deep learning; on the other hand, to verify the conjecture, we propose an approach that generates deep forest holding these characteristics. This is a decision-tree ensemble approach, with fewer hyper-parameters than deep neural networks, and its model complexity can be automatically determined in a data-dependent way. Experiments show that its performance is quite robust to hyper-parameter settings, such that in most cases, even across different data from different domains, it is able to achieve excellent performance by using the same default setting. This study opens the door to deep learning based on non-differentiable modules without gradient-based adjustment, and exhibits the possibility of constructing deep models without backpropagation. Oxford University Press 2019-01 2018-10-08 /pmc/articles/PMC8291612/ /pubmed/34691833 http://dx.doi.org/10.1093/nsr/nwy108 Text en © The Author(s) 2018. Published by Oxford University Press on behalf of China Science Publishing & Media Ltd. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ), which permits non-commercial reuse, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com
spellingShingle Information Science
Zhou, Zhi-Hua
Feng, Ji
Deep forest
title Deep forest
title_full Deep forest
title_fullStr Deep forest
title_full_unstemmed Deep forest
title_short Deep forest
title_sort deep forest
topic Information Science
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8291612/
https://www.ncbi.nlm.nih.gov/pubmed/34691833
http://dx.doi.org/10.1093/nsr/nwy108
work_keys_str_mv AT zhouzhihua deepforest
AT fengji deepforest