Towards Gaussian Bayesian Network Fusion

Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9161)


Data sets are growing in complexity thanks to the increasing facilities we have nowadays to both generate and store data. This poses many challenges to machine learning that are leading to the proposal of new methods and paradigms, in order to be able to deal with what is nowadays referred to as Big Data. In this paper we propose a method for the aggregation of different Bayesian network structures that have been learned from separate data sets, as a first step towards mining data sets that need to be partitioned in an horizontal way, i.e. with respect to the instances, in order to be processed. Considerations that should be taken into account when dealing with this situation are discussed. Scalable learning of Bayesian networks is slowly emerging, and our method constitutes one of the first insights into Gaussian Bayesian network aggregation from different sources. Tested on synthetic data it obtains good results that surpass those from individual learning. Future research will be focused on expanding the method and testing more diverse data sets.


Gaussian Bayesian network Fusion Scalability Big data 



The authors thank the reviewers for comments and critics which significantly contributed to improve the paper; and also J.M. Peña, J. Nielsen, J. Mengin and M. Serrurier for the valuable help. This work has been partially supported by the Spanish Ministry of Economy and Competitiveness through the Cajal Blue Brain (C080020-09; the Spanish partner of the Blue Brain initiative from EPFL) and TIN2013-41592-P projects, and by the Regional Government of Madrid through the S2013/ICE-2845-CASI-CAM-CM project.


  1. 1.
    del Sagrado, J., Moral, S.: Qualitative combination of Bayesian networks. Int. J. Intell. Syst. 18(2), 237–249 (2003)CrossRefzbMATHGoogle Scholar
  2. 2.
    Etminani, K., Naghibzadeh, M., Peña, J.M.: DemocraticOP: A democratic way of aggregating Bayesian network parameters. Int. J. Approximate Reasoning 54(5), 602–614 (2013)MathSciNetCrossRefzbMATHGoogle Scholar
  3. 3.
    Fan, T.H., Lin, D.K., Cheng, K.F.: Regression analysis for massive datasets. Data Knowl. Eng. 61(3), 554–562 (2007)CrossRefGoogle Scholar
  4. 4.
    Geiger, D., Heckerman, D.: Learning Gaussian networks. In: Proceedings of the Tenth International Conference on Uncertainty in Artificial Intelligence, pp. 235–243. Morgan Kaufmann Publishers Inc. (1994)Google Scholar
  5. 5.
    López-Cruz, P.L., Larrañaga, P., DeFelipe, J., Bielza, C.: Bayesian network modeling of the consensus between experts: an application to neuron classification. Int. J. Approximate Reasoning 55(1), 3–22 (2014)MathSciNetCrossRefzbMATHGoogle Scholar
  6. 6.
    Maynard-Reid II, P., Chajewska, U.: Aggregating learned probabilistic beliefs. In: Proceedings of the Eighteenth International Conference on Uncertainty in Artificial Intelligence, pp. 354–361. Morgan Kaufmann Publishers Inc. (2001)Google Scholar
  7. 7.
    Matzkevich, I., Abramson, B.: The topological fusion of Bayes nets. In: Proceedings of the Eighth International Conference on Uncertainty in Artificial Intelligence, pp. 191–198. Morgan Kaufmann Publishers Inc. (1992)Google Scholar
  8. 8.
    Nielsen, J.D., Koc̆ka, T., Peña, J.M.: On local optima in learning Bayesian networks. In: Proceedings of the Nineteenth International Conference on Uncertainty in Artificial Intelligence, pp. 435–442. Morgan Kaufmann Publishers Inc. (2002)Google Scholar
  9. 9.
    Pennock, D.M., Wellman, M.P.: Graphical representation of consensus belief. In: Proceedings of the Fifteenth International Conference on Uncertainty in Artificial Intelligence, pp. 531–540. Morgan Kaufmann Publishers Inc. (1999)Google Scholar
  10. 10.
    Pearl, J.: Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann Publishers Inc, California (1988)zbMATHGoogle Scholar
  11. 11.
    Richardson, M., Domingos, P.: Learning with knowledge from multiple experts. In: Proceedings of the Twentieth International Conference on Machine Learning, pp. 624–631. AAAI Press (2003)Google Scholar
  12. 12.
    Schwarz, G.: Estimating the dimension of a model. Ann. Stat. 6(2), 461–464 (1978)MathSciNetCrossRefzbMATHGoogle Scholar
  13. 13.
    Scutari, M.: Learning Bayesian networks with the bnlearn R package. J. Stat. Softw. 35(3), 1–22 (2010)MathSciNetCrossRefGoogle Scholar
  14. 14.
    Spirtes, P., Glymour, C.N., Scheines, R.: Causation, Prediction, and Search. MIT press, Cambridge (2000)zbMATHGoogle Scholar
  15. 15.
    Tsamardinos, I., Brown, L.E., Aliferis, C.F.: The max-min hill-climbing Bayesian network structure learning algorithm. Mach. Learn. 65(1), 31–78 (2006)CrossRefGoogle Scholar
  16. 16.
    Vidaurre, D., Bielza, C., Larrañaga, P.: Learning an L1-regularized Gaussian Bayesian network in the equivalence class space. IEEE Trans. Syst. Man Cybern. Part B Cybern. 40(5), 1231–1242 (2010)CrossRefGoogle Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  1. 1.Departamento de Inteligencia ArtificialUniversidad Politécnica de MadridMadridSpain

Personalised recommendations