Show pageOld revisionsBacklinksBack to top This page is read only. You can view the source, but not change it. Ask your administrator if you think this is wrong. ====== Expectation Maximization ====== - Iterative approach to MLE when latent variables are present. - Gaussian mixture models approach to density estimation, paramter of distributions are fit using em algorithm. Problem: Estimate joint probability distribution for dataset Density estimation: Select probability distribution and parameters to best explain jpd. Assumption on MLE: All variables that are relevant to the problem are present (no hidden = latent variables). Alternate formulation of maximum likelihood is required for searching appropriate model paramters => EM Algorithm ===== EM Algorithm ===== Two modes: * E-Step: Estimate missing (=latent) variables * M-Step: Maximizte parameters of the model in the presence of the data Usually applied for unsupervised learning (density estimation, clustering). ==== Mixture model ==== Unspecied combination of multiple probability distribution functions. GMM => estimate stddev and mean for each pdf. data_mining/expectation_maximation.txt Last modified: 2022/07/08 21:27by phreazer