Skip to main content

MDS Algorithms

  • Chapter
  • First Online:
  • 1737 Accesses

Part of the book series: SpringerBriefs in Statistics ((BRIEFSSTATIST))

Abstract

Two types of solutions for MDS are discussed. If the proximities are Euclidean distances, classical MDS yields an easy algebraic solution. In most MDS applications, iterative methods are needed, because they admit many types of data and distances. They use a two-phase optimization algorithm, moving the points in MDS space in small steps while holding the data and their transforms fixed, and vice versa, until convergence is reached.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   54.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   69.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    This means that the centroid of the MDS configuration becomes the origin. The coordinates of \(\mathbf {X}\), thus, should sum to 0 in each column of \(\mathbf {X}\). This does not carry any consequences for the distances of \(\mathbf {X}\); that is, any other point could also serve as the origin. However, one point must be picked as an origin to compute scalar products.

  2. 2.

    smacof is an acronym for “Scaling by MAjorizing a COmplicated Function” (De Leeuw and Heiser 1980). The optimization method used by smacof is called “Majorization” (De Leeuw 1977, Groenen 1993). The basic idea of this method is that a complicated goal function (i.e., Stress within the MDS context) is approximated in each iteration by a less complicated function which is easier to optimize. For more details on how this method is used to solve MDS problems, see De Leeuw and Mair (2009) or Borg and Groenen (2005).

  3. 3.

    Such options are sometimes called Kruskal, Guttman, Young or Torgerson, depending on their respective inventors or authors (see also Fig. 1.5 and Fig. 10.4).

References

  • Basalaj, W. (2001). Proximity visualisation of abstract data, Unpublished doctoral dissertation, Cambridge University, U.K.

    Google Scholar 

  • Borg, I., & Groenen, P. J. F. (2005). Modern multidimensional scaling (2nd ed.). New York: Springer.

    MATH  Google Scholar 

  • De Leeuw, J. (1977). Applications of convex analysis to multidimensional scaling. In J. R. Barra, F. Brodeau, G. Romier, & B. van Cutsem (Eds.), Recent developments in statistics (pp. 133–145). Amsterdam: North Holland.

    Google Scholar 

  • De Leeuw, J., & Mair, P. (2009). Multidimensional scaling using majorization: SMACOF in R. Journal of Statistical Software, 31(3), 1-30. Retrieved from http://www.jstatsoft.org/v31/i03/.

  • De Leeuw, J., & Heiser, W. J. (1980). Multidimensional scaling with restrictions on the configuration. In P. R. Krishnaiah (Ed.), Multivariate Analysis (Vol. V, pp. 501–522). Amsterdam: North-Holland.

    Google Scholar 

  • Groenen, P. J. F. (1993). The majorization approach to multidimensional scaling: some problems and extensions, Unpublished doctoral dissertation, University of Leiden.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ingwer Borg .

Rights and permissions

Reprints and permissions

Copyright information

© 2018 The Author(s)

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Borg, I., Groenen, P.J.F., Mair, P. (2018). MDS Algorithms. In: Applied Multidimensional Scaling and Unfolding. SpringerBriefs in Statistics. Springer, Cham. https://doi.org/10.1007/978-3-319-73471-2_9

Download citation

Publish with us

Policies and ethics