Cargando…

Error Bound of Mode-Based Additive Models

Due to their flexibility and interpretability, additive models are powerful tools for high-dimensional mean regression and variable selection. However, the least-squares loss-based mean regression models suffer from sensitivity to non-Gaussian noises, and there is also a need to improve the model’s...

Descripción completa

Detalles Bibliográficos
Autores principales: Deng, Hao, Chen, Jianghong, Song, Biqin, Pan, Zhibin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8224641/
https://www.ncbi.nlm.nih.gov/pubmed/34067420
http://dx.doi.org/10.3390/e23060651
_version_ 1783711930433667072
author Deng, Hao
Chen, Jianghong
Song, Biqin
Pan, Zhibin
author_facet Deng, Hao
Chen, Jianghong
Song, Biqin
Pan, Zhibin
author_sort Deng, Hao
collection PubMed
description Due to their flexibility and interpretability, additive models are powerful tools for high-dimensional mean regression and variable selection. However, the least-squares loss-based mean regression models suffer from sensitivity to non-Gaussian noises, and there is also a need to improve the model’s robustness. This paper considers the estimation and variable selection via modal regression in reproducing kernel Hilbert spaces (RKHSs). Based on the mode-induced metric and two-fold Lasso-type regularizer, we proposed a sparse modal regression algorithm and gave the excess generalization error. The experimental results demonstrated the effectiveness of the proposed model.
format Online
Article
Text
id pubmed-8224641
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-82246412021-06-25 Error Bound of Mode-Based Additive Models Deng, Hao Chen, Jianghong Song, Biqin Pan, Zhibin Entropy (Basel) Article Due to their flexibility and interpretability, additive models are powerful tools for high-dimensional mean regression and variable selection. However, the least-squares loss-based mean regression models suffer from sensitivity to non-Gaussian noises, and there is also a need to improve the model’s robustness. This paper considers the estimation and variable selection via modal regression in reproducing kernel Hilbert spaces (RKHSs). Based on the mode-induced metric and two-fold Lasso-type regularizer, we proposed a sparse modal regression algorithm and gave the excess generalization error. The experimental results demonstrated the effectiveness of the proposed model. MDPI 2021-05-22 /pmc/articles/PMC8224641/ /pubmed/34067420 http://dx.doi.org/10.3390/e23060651 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Deng, Hao
Chen, Jianghong
Song, Biqin
Pan, Zhibin
Error Bound of Mode-Based Additive Models
title Error Bound of Mode-Based Additive Models
title_full Error Bound of Mode-Based Additive Models
title_fullStr Error Bound of Mode-Based Additive Models
title_full_unstemmed Error Bound of Mode-Based Additive Models
title_short Error Bound of Mode-Based Additive Models
title_sort error bound of mode-based additive models
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8224641/
https://www.ncbi.nlm.nih.gov/pubmed/34067420
http://dx.doi.org/10.3390/e23060651
work_keys_str_mv AT denghao errorboundofmodebasedadditivemodels
AT chenjianghong errorboundofmodebasedadditivemodels
AT songbiqin errorboundofmodebasedadditivemodels
AT panzhibin errorboundofmodebasedadditivemodels