z-logo
Premium
ROBUST RIDGE REGRESSION BASED ON AN M‐ESTIMATOR
Author(s) -
SILVAPULLE MERVYN J.
Publication year - 1991
Publication title -
australian journal of statistics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.434
H-Index - 41
eISSN - 1467-842X
pISSN - 0004-9581
DOI - 10.1111/j.1467-842x.1991.tb00438.x
Subject(s) - estimator , mathematics , ordinary least squares , mean squared error , ridge , efficient estimator , invariant estimator , trimmed estimator , robust regression , bias of an estimator , minimum variance unbiased estimator , outlier , statistics , consistent estimator , james–stein estimator , least squares function approximation , geology , paleontology
Summary Consider the linear regression model y =β 0 1 + Xβ + in the usual notation. It is argued that the class of ordinary ridge estimators obtained by shrinking the least squares estimator by the matrix (X 1 X + kI) ‐1 X'X is sensitive to outliers in the ^variable. To overcome this problem, we propose a new class of ridge‐type M‐estimators, obtained by shrinking an M‐estimator (instead of the least squares estimator) by the same matrix. Since the optimal value of the ridge parameter k is unknown, we suggest a procedure for choosing it adaptively. In a reasonably large scale simulation study with a particular M‐estimator, we found that if the conditions are such that the M‐estimator is more efficient than the least squares estimator then the corresponding ridge‐type M‐estimator proposed here is better, in terms of a Mean Squared Error criteria, than the ordinary ridge estimator with k chosen suitably. An example illustrates that the estimators proposed here are less sensitive to outliers in the y‐variable than ordinary ridge estimators.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here