Full text
![]() |
PDF
- Requires a PDF viewer, such as GSview, Xpdf or Adobe Acrobat Reader
Download (834kB) |
Vidaurre Henche, Diego and Bielza Lozoya, María Concepción and Larrañaga Múgica, Pedro María (2013). An L1-Regularized naïve bayes-inspired classifier for discarding redundant and irrelevant predictors. "International Journal on Artificial Intelligence Tools", v. 22 (n. 4); pp.. ISSN 1793-6349. https://doi.org/10.1142/S021821301350019X.
Title: | An L1-Regularized naïve bayes-inspired classifier for discarding redundant and irrelevant predictors |
---|---|
Author/s: |
|
Item Type: | Article |
Título de Revista/Publicación: | International Journal on Artificial Intelligence Tools |
Date: | August 2013 |
ISSN: | 1793-6349 |
Volume: | 22 |
Subjects: | |
Freetext Keywords: | Lasso, Regularization, Naïve Bayes, Redundancy |
Faculty: | E.T.S. de Ingenieros Informáticos (UPM) |
Department: | Inteligencia Artificial |
Creative Commons Licenses: | Recognition - No derivative works - Non commercial |
![]() |
PDF
- Requires a PDF viewer, such as GSview, Xpdf or Adobe Acrobat Reader
Download (834kB) |
The naïve Bayes model is a simple but often satisfactory supervised classification method. The original naïve Bayes scheme, does, however, have a serious weakness, namely, the harmful effect of redundant predictors. In this paper, we study how to apply a regularization technique to learn a computationally efficient classifier that is inspired by naïve Bayes. The proposed formulation, combined with an L1-penalty, is capable of discarding harmful, redundant predictors. A modification of the LARS algorithm is devised to solve this problem. We tackle both real-valued and discrete predictors, assuring that our method is applicable to a wide range of data. In the experimental section, we empirically study the effect of redundant and irrelevant predictors. We also test the method on a high dimensional data set from the neuroscience field, where there are many more predictors than data cases. Finally, we run the method on a real data set than combines categorical with numeric predictors. Our approach is compared with several naïve Bayes variants and other classification algorithms (SVM and kNN), and is shown to be competitive.
Item ID: | 72794 |
---|---|
DC Identifier: | https://oa.upm.es/72794/ |
OAI Identifier: | oai:oa.upm.es:72794 |
DOI: | 10.1142/S021821301350019X |
Official URL: | https://www.worldscientific.com/doi/10.1142/S02182... |
Deposited by: | Biblioteca Facultad de Informatica |
Deposited on: | 17 Mar 2023 11:34 |
Last Modified: | 17 Mar 2023 11:34 |