Please use this identifier to cite or link to this item:
|Title:||Estimating local optimums in EM algorithm over Gaussian mixture model|
|Authors:||Zhang, Z. |
|Citation:||Zhang, Z.,Bing, T.D.,Tung, A.K.H. (2008). Estimating local optimums in EM algorithm over Gaussian mixture model. Proceedings of the 25th International Conference on Machine Learning : 1240-1247. ScholarBank@NUS Repository.|
|Abstract:||EM algorithm is a very popular iteration-based method to estimate the parameters of Gaussian Mixture Model from a large observation set. However, in most cases, EM algorithm is not guaranteed to converge to the global optimum. Instead, it stops at some local optimums, which can be much worse than the global optimum. Therefore, it is usually required to run multiple procedures of EM algorithm with different initial configurations and return the best solution. To improve the efficiency of this scheme, we propose a new method which can estimate an upper bound on the logarithm likelihood of the local optimum, based on the current configuration after the latest EM iteration. This is accomplished by first deriving some region bounding the possible locations of local optimum, followed by some upper bound estimation on the maximum likelihood. With this estimation, we can terminate an EM algorithm procedure if the estimated local optimum is definitely worse than the best solution seen so far. Extensive experiments show that our method can effectively and efficiently accelerate conventional multiple restart EM algorithm. Copyright 2008 by the author(s)/owner(s).|
|Source Title:||Proceedings of the 25th International Conference on Machine Learning|
|Appears in Collections:||Staff Publications|
Show full item record
Files in This Item:
There are no files associated with this item.
checked on Nov 17, 2018
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.