Computational aspects of fitting mixture models via the expectation–maximization algorithm
O’Hagan, Adrian,
Thomas Brendan Murphy and
Isobel Claire Gormley
Computational Statistics & Data Analysis, 2012, vol. 56, issue 12, 3843-3864
Abstract:
The Expectation–Maximization (EM) algorithm is a popular tool in a wide variety of statistical settings, in particular in the maximum likelihood estimation of parameters when clustering using mixture models. A serious pitfall is that in the case of a multimodal likelihood function the algorithm may become trapped at a local maximum, resulting in an inferior clustering solution. In addition, convergence to an optimal solution can be very slow. Methods are proposed to address these issues: optimizing starting values for the algorithm and targeting maximization steps efficiently. It is demonstrated that these approaches can produce superior outcomes to initialization via random starts or hierarchical clustering and that the rate of convergence to an optimal solution can be greatly improved.
Keywords: Convergence rate; Expectation–maximization algorithm; Hierarchical clustering; mclust; Model-based clustering; Multimodal likelihood (search for similar items in EconPapers)
Date: 2012
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (7)
Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/S0167947312002101
Full text for ScienceDirect subscribers only.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:eee:csdana:v:56:y:2012:i:12:p:3843-3864
DOI: 10.1016/j.csda.2012.05.011
Access Statistics for this article
Computational Statistics & Data Analysis is currently edited by S.P. Azen
More articles in Computational Statistics & Data Analysis from Elsevier
Bibliographic data for series maintained by Catherine Liu ().