Advertisement

On Identifying Tree-Structured Perfect Maps

  • Christian Borgelt
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 2821)

Abstract

It is well known that tree-structured perfect maps can be uniquely identified by computing a maximum weight spanning tree with mutual information providing the edge weights. In this paper I generalize the edge evaluation measure by stating the conditions such a measure has to satisfy in order to be able to identify tree-structured perfect maps. In addition, I show that not only mutual information, but also the well-known χ 2 measure satisfies these conditions.

Keywords

Mutual Information Edge Weight Conditional Independence Joint Probability Distribution Markov Network 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Borgelt, C., Kruse, R.: Graphical Models — Methods for Data Analysis and Mining. J. Wiley & Sons, Chichester (2002)zbMATHGoogle Scholar
  2. 2.
    Castillo, E., Gutierrez, J.M., Hadi, A.S.: Expert Systems and Probabilistic Network Models. Springer, New York (1997)Google Scholar
  3. 3.
    Chow, C.K., Liu, C.N.: Approximating Discrete Probability Distributions with Dependence Trees. IEEE Trans. on Information Theory 14(3), 462–467 (1968)zbMATHCrossRefMathSciNetGoogle Scholar
  4. 4.
    Friedman, N., Goldszmidt, M.: Building Classifiers using Bayesian Networks. In: Proc. 13th Nat. Conf. on Artificial Intelligence, AAAI 1996, Portland, OR, USA, pp. 1277–1284. AAAI Press, Menlo Park (1996)Google Scholar
  5. 5.
    Geiger, D.: An entropy-based learning algorithm of Bayesian conditional trees. In: Proc. 8th Conf. on Uncertainty in Artificial Intelligence, UAI 1992, Stanford, CA, USA, pp. 92–97. Morgan Kaufmann, San Mateo (1992)Google Scholar
  6. 6.
    Jensen, F.V.: An Introduction to Bayesian Networks. UCL Press, London (1996)Google Scholar
  7. 7.
    Kruskal, J.B.: On the Shortest Spanning Subtree of a Graph and the Traveling Salesman Problem. In: Proc. American Mathematical Society, vol. 7(1), pp. 48–50. American Mathematical Society, Providence (1956)Google Scholar
  8. 8.
    Kullback, S., Leibler, R.A.: On Information and Sufficiency. Annals of Mathematical Statistics 22, 79–86 (1951)zbMATHCrossRefMathSciNetGoogle Scholar
  9. 9.
    Lauritzen, S.L.: Graphical Models. Oxford University Press, Oxford (1996)Google Scholar
  10. 10.
    Pearl, J.: Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, San Mateo (1988) (2nd edition 1992)Google Scholar
  11. 11.
    Press, W.H., Teukolsky, S.A., Vetterling, W.T., Flannery, B.P.: Numerical Recipes in C — The Art of Scientific Computing, 2nd edn. Cambridge University Press, Cambridge (1992)zbMATHGoogle Scholar
  12. 12.
    Prim, R.C.: Shortest Connection Networks and Some Generalizations. The Bell System Technical Journal 36, 1389–1401 (1957)Google Scholar
  13. 13.
    Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)Google Scholar
  14. 14.
    Rebane, G., Pearl, J.: The Recovery of Causal Polytrees from Statistical Data. In: Proc. 3rd Workshop on Uncertainty in Artificial Intelligence, Seattle, WA, USA, pp. 222–228 (1987)Google Scholar
  15. 15.
    Shannon, C.E.: The Mathematical Theory of Communication. The Bell System Technical Journal 27, 379–423 (1948)zbMATHMathSciNetGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2003

Authors and Affiliations

  • Christian Borgelt
    • 1
  1. 1.Dept. of Knowledge Processing and Language EngineeringSchool of Computer Science, Otto-von-Guericke-University of MagdeburgMagdeburgGermany

Personalised recommendations