Skip to main content

The EM Algorithm

  • Chapter
  • First Online:
Optimization

Part of the book series: Springer Texts in Statistics ((STS,volume 95))

Abstract

Maximum likelihood is the dominant form of estimation in applied statistics. Because closed-form solutions to likelihood equations are the exception rather than the rule, numerical methods for finding maximum likelihood estimates are of paramount importance. In this chapter we study maximum likelihood estimation by the EM algorithm a special case of the MM algorithm. At the heart of every EM algorithm is some notion of missing data. Data can be missing in the ordinary sense of a failure to record certain observations on certain cases. Data can also be missing in a theoretical sense. We can think of the E (expectation) step of the algorithm as filling in the missing data. This action replaces the loglikelihood of the observed data by a minorizing function. This surrogate function is then maximized in the M step. Because the surrogate function is usually much simpler than the likelihood, we can often solve the M step analytically. The price we pay for this simplification is that the EM algorithm is iterative. Reconstructing the missing data is bound to be slightly wrong if the parameters do not already equal their maximum likelihood estimates.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 179.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Baum LE (1972) An inequality and associated maximization technique in statistical estimation for probabilistic functions of Markov processes. Inequalities 3:1–8

    Google Scholar 

  2. Dempster AP, Laird NM, Rubin DB (1977) Maximum likelihood from incomplete data via the EM algorithm (with discussion). J Roy Stat Soc B 39:1–38

    MathSciNet  MATH  Google Scholar 

  3. Devijver PA (1985) Baum’s forward-backward algorithm revisited. Pattern Recogn Lett 3:369–373

    Article  MATH  Google Scholar 

  4. Dobson AJ (1990) An introduction to generalized linear models. Chapman & Hall, London

    MATH  Google Scholar 

  5. Duan J-C, Simonato J-G (1993) Multiplicity of solutions in maximum likelihood factor analysis. J Stat Comput Simul 47:37–47

    Article  Google Scholar 

  6. Durbin R, Eddy S, Krogh A, Mitchison G (1998) Biological sequence analysis: probabilistic models of proteins and nucleic acids. Cambridge University Press, Cambridge

    Book  MATH  Google Scholar 

  7. Flury B, Zoppè A (2000) Exercises in EM. Am Stat 54:207–209

    Google Scholar 

  8. Green PJ (1990) On use of the EM algorithm for penalized likelihood estimation. J Roy Stat Soc B 52:443–452

    MATH  Google Scholar 

  9. Lange K (2002) Mathematical and statistical methods for genetic analysis, 2nd edn. Springer, New York

    Book  MATH  Google Scholar 

  10. Lange K (2010) Numerical analysis for statisticians, 2nd edn. Springer, New York

    Book  MATH  Google Scholar 

  11. Little RJA, Rubin DB (1987) Statistical analysis with missing data. Wiley, Hoboken

    MATH  Google Scholar 

  12. McLachlan GJ, Krishnan T (2008) The EM algorithm and extensions, 2nd edn. Wiley, Hoboken

    Book  MATH  Google Scholar 

  13. McLachlan GJ, Peel D (2000) Finite mixture models. Wiley, Hoboken

    Book  MATH  Google Scholar 

  14. Rabiner L (1989) A tutorial on hidden Markov models and selected applications in speech recognition. Proc IEEE 77:257–285

    Article  Google Scholar 

  15. Rao CR (1973) Linear statistical inference and its applications, 2nd edn. Wiley, Hoboken

    Book  MATH  Google Scholar 

  16. Tanner MA (1993) Tools for statistical inference: methods for the exploration of posterior distributions and likelihood functions, 2nd edn. Springer, New York

    MATH  Google Scholar 

  17. Titterington DM, Smith AFM, Makov UE (1985) Statistical analysis of finite mixture distributions. Wiley, Hoboken

    MATH  Google Scholar 

  18. Weeks DE, Lange K (1989) Trials, tribulations, and triumphs of the EM algorithm in pedigree analysis. IMA J Math Appl Med Biol 6:209–232

    Article  MathSciNet  MATH  Google Scholar 

  19. Zhou H, Lange K (2009) On the bumpy road to the dominant mode. Scand J Stat 37:612–631

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer Science+Business Media New York

About this chapter

Cite this chapter

Lange, K. (2013). The EM Algorithm. In: Optimization. Springer Texts in Statistics, vol 95. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-5838-8_9

Download citation

Publish with us

Policies and ethics