z-logo
Premium
Maximum likelihood estimation of a multi‐dimensional log‐concave density
Author(s) -
Cule Madeleine,
Samworth Richard,
Stewart Michael
Publication year - 2010
Publication title -
journal of the royal statistical society: series b (statistical methodology)
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 6.523
H-Index - 137
eISSN - 1467-9868
pISSN - 1369-7412
DOI - 10.1111/j.1467-9868.2010.00753.x
Subject(s) - mathematics , estimator , kernel density estimation , smoothing , density estimation , independent and identically distributed random variables , algorithm , multivariate kernel density estimation , variable kernel density estimation , mathematical optimization , kernel (algebra) , combinatorics , random variable , kernel method , statistics , computer science , artificial intelligence , support vector machine
Summary.  Let X 1 ,…, X n be independent and identically distributed random vectors with a (Lebesgue) density f . We first prove that, with probability 1, there is a unique log‐concave maximum likelihood estimator of f . The use of this estimator is attractive because, unlike kernel density estimation, the method is fully automatic, with no smoothing parameters to choose. Although the existence proof is non‐constructive, we can reformulate the issue of computing in terms of a non‐differentiable convex optimization problem, and thus combine techniques of computational geometry with Shor's r ‐algorithm to produce a sequence that converges to . An R version of the algorithm is available in the package LogConcDEAD—log‐concave density estimation in arbitrary dimensions. We demonstrate that the estimator has attractive theoretical properties both when the true density is log‐concave and when this model is misspecified. For the moderate or large sample sizes in our simulations, is shown to have smaller mean integrated squared error compared with kernel‐based methods, even when we allow the use of a theoretical, optimal fixed bandwidth for the kernel estimator that would not be available in practice. We also present a real data clustering example, which shows that our methodology can be used in conjunction with the expectation–maximization algorithm to fit finite mixtures of log‐concave densities.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here