Abstract
As an extremely powerful probability model, Gaussian mixture model (GMM) has been widely used in fields of pattern recognition, information processing and data mining. If the number of the Gaussians in the mixture is pre-known, the well-known Expectation-Maximization (EM) algorithm could be used to estimate the parameters in the Gaussian mixture model. However, in many practical applications, the number of the components is not known.Then the Gaussian mixture modeling becomes a compound problem of the determination of number of Gaussian components and the parameter estimation for the mixture, which is rather difficult. In this paper, we propose a split and merge EM (SMEM) algorithm to decide the number of the components, which is referred to the model selection for the mixture. Based on the minimum description length (MDL) criterion, the proposed SMEM algorithm can avoid the local optimum drawback of the usual EM algorithm and determine the number of components in the Gaussian mixture model automatically. By splitting and merging the uncorrect components, the algorithm can converge to the maximization of the MDL criterion function and get a better parameter estimation of the Gaussian mixture with correct number of components in the mixture. It is demonstrated well by the experiments that the proposed split and merge EM algorithm can make both parameter learning and model selection efficiently for Gaussian mixture.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.