Applying Bayesian Networks for Meteorological Data Mining

  • Estevam R. HruschkaJr
  • Eduardo R. Hruschka
  • Nelson F. F. Ebecken


Bayesian Networks (BNs) have been recently employed to solve meteorology problems. In this paper, the application of BNs for mining a real-world weather dataset is described. The employed dataset discriminates between “wet fog” instances and “other weather conditions” instances, and it contains many missing data. Therefore, BNs were employed not only for classifying instances, but also for filling missing data. In addition, the Markov Blanket concept was employed to select relevant attributes. The efficacy of BNs to perform the aforementioned tasks was assessed by means of several experiments. In summary, more convincing results were obtained by taking advantage of the fact that BNs can directly (i.e. without data preparation) classify instances containing missing values. In addition, the attributes selected by means of the Markov Blanket provide a simpler, faster, and equally accurate classifier.


Data Mining Bayesian Network Conditional Independence Data Preparation Bayesian Classifier 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Basak, J., Sudarshan, A., Trivedi, D., Santhanam, M.S., Weather Data Mining Using Independent Component Analysis, Journal of Machine Learning Research, n.5, pp. 239–253, 2004.MathSciNetGoogle Scholar
  2. 2.
    Cano, R., Sordo, C., Gutiérrez, J.M., Applications of Bayesian Networks in Meteorology, Advances in Bayesian Networks, Gámez, J.A. et al. eds., pp. 309–327, Springer, 2004.Google Scholar
  3. 3.
    Cofiño, A.S., Gutiérrez, J.M., Jakubiak, B., Melonek, M., Implementation of data mining techniques for meteorological applications. In: Realizing Teracomputing, Zwieflhofer, W. & N. Kreitz eds., pp. 256–271, World Scientific Publishing, 2003.Google Scholar
  4. 4.
    Heckerman, D. “Bayesian networks for data mining,” Data Mining and Knowledge Discovery, vol. 1, pp. 79–119, 1997.CrossRefGoogle Scholar
  5. 5.
    Hruschka JR., E. R., Hruschka, E. R., Ebecken, N. F. F. A Data Preparation Bayesian Approach for a Clustering Genetic Algorithm. In: Frontiers in Artificial Intelligence and Applications, Soft Computing Systems: Design, Management and Applications, IOS Press, v.87, pp. 453–461, 2002.Google Scholar
  6. 6.
    Blum, A.L., Langley, P., Selection of Relevant Features and Examples in Machine Learning, Artificial Intelligence, pp. 245–271, 1997.Google Scholar
  7. 7.
    Hruschka JR., E. R., Hruschka, E. R., Ebecken, N. F. F. Feature Selection by Bayesian Networks In: The Seventeenth Canadian Conference on Artificial Intelligence, 2004, London, Ontario. Lecture Notes in Artificial Intelligence. Berlin: Springer-Verlag, v. 3060, pp. 370–379, 2004.Google Scholar
  8. 8.
    Pearl, J., Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference, Morgan Kaufmann, San Mateo, CA, 1988.Google Scholar
  9. 9.
    Friedman, N. and Koller, D., Being Bayesian about network structure. A Bayesian approach to structure discovery in Bayesian networks, Machine Leraning 50(1–2): 95–125, 2003.zbMATHCrossRefGoogle Scholar
  10. 10.
    Cooper, Gregory F. NESTOR: A computer-based medical diagnostic aid that integrates causal and probabilistic knowledge, PhD. thesis, Rep. No. STAN-CS-84-48 (also HPP-84-48) Dept. of Computer Science, Stanford Univ., CA, 1984.Google Scholar
  11. 11.
    Chickering, D. M., Optimal Structure Identification with Greedy Search, Journal of Machine Learning Research, (3):507–554, 2002.CrossRefMathSciNetGoogle Scholar
  12. 12.
    Spirtes, P., Glymour, C. and Scheines, R., Causation, Prediction, and Search, (Adaptive Computation and Machine Learning), 2nd edition, Bradford Books, 2001.Google Scholar
  13. 13.
    Cheng, J., Greiner, R., Kelly, J., Bell, D., Liu, W.R., Learning Bayesian networks from data: An information-theory based approach. Artificial Intelligence, 137(1–2): 43–90, 2002.zbMATHCrossRefMathSciNetGoogle Scholar
  14. 14.
    Cooper G. & Herskovitz, E.. A Bayesian Method for the Induction of Probabilistic Networks from Data. Machine Learning, 9, 309–347, 1992.zbMATHGoogle Scholar
  15. 15.
    Langley, P. & Sage, S., Induction of Selective Bayesian Classifiers. Proceedings of the Tenth Conference on Uncertainty in Artificial Intelligence, Seattle, 1994.Google Scholar
  16. 16.
    Anderson, J. R. & Matessa, M., Explorations of an Incremental Bayesian Algorithm for categorization. Machine Learning, 9, 275–308, 1992.Google Scholar
  17. 17.
    Hsu, W. H., Genetic Wrappers for feature selection in decision tree induction and variable ordering in Bayesian network structure learning, Information Science, 163, pp. 103–122,2004.CrossRefGoogle Scholar
  18. 18.
    Cheng, J. and Greiner, R., Comparing Bayesian Network Classifiers, Proc. of the Fifteenth Conference on Uncertainty in Artificial Intelligence (UAI’ 99), Sweden, pp. 101–108, 1999.Google Scholar
  19. 19.
    Ying, Y. and Webb, G., On Why Discretization Works for Naive-Bayes Classifiers. In Proceedings of the 16th Australian Conference on AI (AI 03), Lecture Notes AI 2903, 440–452. Berlin: Springer, 2003.Google Scholar
  20. 20.
    Ying, Y., Discretization for Naive-Bayes Learning. PhD. Thesis, Monash University, 2003b. Scholar
  21. 21.
    Witten, I. H., Frank, E., Data Mining — Practical Machine Learning Tools and Techniques with Java Implementations, Morgan Kaufmann Publishers, USA, 2000.Google Scholar
  22. 22.
    Dempster, A. P., Laird, N. M., Rubin, D. B., Maximum Likelihood from Incomplete Data via the EM algorithm, Journal of the Royal Statistical Society B, 39,1–39, 1977.Google Scholar
  23. 23.
    Gelfand, A.,E. and Smith, A. F. M., Sampling-based approaches to calculating marginal densities. J. American Statistical Association, 85:398–409, 1990.zbMATHCrossRefMathSciNetGoogle Scholar
  24. 24.
    Casella, G. and George, E. I., “Explaining the Gibbs sampler,” Amer. Statist., vol. 46, pp. 167–174, 1992.CrossRefMathSciNetGoogle Scholar
  25. 25.
    Bigus, J. P., Data Mining with Neural Networks, First edition, USA, McGraw-Hill, 1996.Google Scholar
  26. 26.
    Han, J. and Kamber, M., Data Mining, Concepts and Techniques. Morgan Kaufmann, 2001.Google Scholar
  27. 27.
    Reunanen, J., Overfitting in Making Comparissons Between Variable Selection Methods, Journal of Machine Learning Research 3, pp. 1371–1382, 2003.zbMATHCrossRefGoogle Scholar
  28. 28.
    Liu, H. and Motoda, H., Feature Selection for Knowledge Discovery and Data Mining. Kluwer Academic, 1998.Google Scholar
  29. 29.
    Guyon, I., Elisseeff, A., An Introduction to Variable and Feature Selection, Journal of Machine Learning Research 3, pp. 1157–1182, 2003.zbMATHCrossRefGoogle Scholar
  30. 30.
    Little, R., & Rubin, D. B., Statistical Analysis with Missing Data. Wiley, New York, 1987.zbMATHGoogle Scholar
  31. 31.
    Lauritzen, S. L., & Spiegelhalter, D. J., Local computations with probabilities on graphical structures and their application to expert systems. J. Royal Statistical Society B, 50, 157–224, 1988.zbMATHMathSciNetGoogle Scholar

Copyright information

© Springer-Verlag London Limited 2006

Authors and Affiliations

  • Estevam R. HruschkaJr
    • 1
  • Eduardo R. Hruschka
    • 2
  • Nelson F. F. Ebecken
    • 3
  1. 1.DC-UFSCar/ Federal University of Sao CarlosBrazil
  2. 2.Catholic University of Santos (UniSantos)Brazil
  3. 3.COPPE/ Federal University of Rio de JaneiroBrazil

Personalised recommendations