Premium
Robust penalized logistic regression with truncated loss functions
Author(s) -
Park Seo Young,
Liu Yufeng
Publication year - 2011
Publication title -
canadian journal of statistics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.804
H-Index - 51
eISSN - 1708-945X
pISSN - 0319-5724
DOI - 10.1002/cjs.10105
Subject(s) - outlier , logistic regression , computer science , robust regression , regression , statistics , selection (genetic algorithm) , logistic model tree , mathematics , artificial intelligence , function (biology) , robustness (evolution) , machine learning , pattern recognition (psychology) , econometrics , biochemistry , chemistry , evolutionary biology , gene , biology
The penalized logistic regression (PLR) is a powerful statistical tool for classification. It has been commonly used in many practical problems. Despite its success, since the loss function of the PLR is unbounded, resulting classifiers can be sensitive to outliers. To build more robust classifiers, we propose the robust PLR (RPLR) which uses truncated logistic loss functions, and suggest three schemes to estimate conditional class probabilities. Connections of the RPLR with some other existing work on robust logistic regression have been discussed. Our theoretical results indicate that the RPLR is Fisher consistent and more robust to outliers. Moreover, we develop estimated generalized approximate cross validation (EGACV) for the tuning parameter selection. Through numerical examples, we demonstrate that truncating the loss function indeed yields better performance in terms of classification accuracy and class probability estimation. The Canadian Journal of Statistics 39: 300–323; 2011 © 2011 Statistical Society of Canada