Skip to main content
Log in

Effective sparsity control in deep belief networks using normal regularization term

  • Regular Paper
  • Published:
Knowledge and Information Systems Aims and scope Submit manuscript

Abstract

Nowadays the use of deep network architectures has become widespread in machine learning. Deep belief networks (DBNs) have deep network architectures to create a powerful generative model using training data. Deep belief networks can be used in classification and feature learning. A DBN can be learned unsupervised, and then the learned features are suitable for a simple classifier (like a linear classifier) with a few labeled data. In addition, according to researches, by using sparsity in DBNs we can learn useful low-level feature representations for unlabeled data. In sparse representation, we have the property that learned features can be interpreted, i.e., correspond to meaningful aspects of the input, and capture factors of variation in the data. Different methods are proposed to build sparse DBNs. In this paper, we proposed a new method that has different behavior according to deviation of the activation of the hidden units from a (low) fixed value. In addition, our proposed regularization term has a variance parameter that can control the force degree of sparseness. According to the results, our new method achieves the best recognition accuracy on the test sets in different datasets with different applications (image, speech and text) and we can achieve incredible results when using a different number of training samples, especially when we have a few samples for training.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Notes

  1. Available online at “http://ceit.aut.ac.ir/~keyvanrad/DeeBNetToolbox.html”.

  2. Available online at “http://yann.lecun.com/exdb/mnist/”.

  3. Available online at “https://archive.ics.uci.edu/ml/datasets/ISOLET”.

  4. Available online at “http://qwone.com/~jason/20Newsgroups”.

  5. Available online at “http://qwone.com/~jason/20Newsgroups/20news-bydate-matlab.tgz”.

References

  1. Bai L, Li K, Pei J, Jiang S (2015) Main objects interaction activity recognition in real images. Neural Comput Appl 27:335–348. doi:10.1007/s00521-015-1846-7

    Article  Google Scholar 

  2. Bengio Y (2009) Learning deep architectures for AI. Found Trends Mach Learn 2:1–127

    Article  MATH  Google Scholar 

  3. Fanty M, Cole RA (1990) Spoken letter recognition. In: Lippmann RP, Moody JE, Touretzky DS (eds) Proceedings of the 3rd international conference on neural information processing systems (NIPS’90). Morgan Kaufmann Publishers Inc., San Francisco, CA, pp 220–226

  4. Fischer A, Igel C (2014) Training restricted Boltzmann machines: an introduction. Pattern Recognit 47:25–39. doi:10.1016/j.patcog.2013.05.025

    Article  MATH  Google Scholar 

  5. Halkias X, Paris S, Glotin H (2013) Sparse penalty in deep belief networks: using the mixed norm constraint. arXiv:1301.3533 [cs, stat]

  6. Hinton G (2010) A practical guide to training restricted Boltzmann machines. Machine Learning Group, University of Toronto

  7. Hinton GE, Osindero S, Teh Y-W (2006) A fast learning algorithm for deep belief nets. Neural Comput 18:1527–1554

    Article  MathSciNet  MATH  Google Scholar 

  8. Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313:504–507

    Article  MathSciNet  MATH  Google Scholar 

  9. Ji N-N, Zhang J-S, Zhang C-X (2014) A sparse-response deep belief network based on rate distortion theory. Pattern Recognit 47:3179–3191

    Article  Google Scholar 

  10. Keyvanrad MA, Homayounpour MM (2015a) Normal sparse deep belief network. In: 2015 international joint conference on neural networks (IJCNN), pp 1–7

  11. Keyvanrad MA, Homayounpour MM (2015b) Deep belief network training improvement using elite samples minimizing free energy. Int J Patt Pattern Artif Intell. doi:10.1142/S0218001415510064

    Google Scholar 

  12. Keyvanrad MA, Homayounpour MM (2014) A brief survey on deep belief networks and introducing a new object oriented toolbox (DeeBNet). arXiv:1408.3264 [cs]

  13. LeCun Y, Cortes C, Burges CJ (1998) The MNIST database of handwritten digits. The MNIST data set is available at http://yann.lecun.com/exdb/mnist/index.html

  14. Lee H (2010) Unsupervised feature learning via sparse hierarchical representations. Ph.D. Thesis, Stanford University

  15. Lee H, Ekanadham C, Ng A (2008) Sparse deep belief net model for visual area V2. Adv Neural Inf Process Syst 20:873–880

    Google Scholar 

  16. Liu Y, Zhou S, Chen Q (2011) Discriminative deep belief networks for visual data classification. Pattern Recognit 44:2287–2296

    Article  MATH  Google Scholar 

  17. Marc’Aurelio Ranzato Y, Boureau L, LeCun Y (2007) Sparse feature learning for deep belief networks. Adv Neural Inf Process Syst 20:1185–1192

    Google Scholar 

  18. Mohamed A, Dahl G, Hinton G (2009) Deep belief networks for phone recognition. In: NIPS workshop on deep learning for speech recognition and related applications. Canada, pp 1–9

  19. Murphy KP (2012) Machine learning: a probabilistic perspective. The MIT Press, Cambridge

    MATH  Google Scholar 

  20. Nair V (2010) Visual object recognition using generative models of images. Ph.D Thesis, University of Toronto

  21. Nair V, Hinton G (2009) 3D object recognition with deep belief nets. Adv Neural Inf Process Syst 22:1339–1347

    Google Scholar 

  22. Olshausen BA, David JF (1996) Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 381:607–609

    Article  Google Scholar 

  23. Tieleman T (2008) Training restricted Boltzmann machines using approximations to the likelihood gradient. In: Proceedings of the 25th international conference on Machine learning. ACM, New York, NY, USA, pp 1064–1071

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohammad Mehdi Homayounpour.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Keyvanrad, M.A., Homayounpour, M.M. Effective sparsity control in deep belief networks using normal regularization term. Knowl Inf Syst 53, 533–550 (2017). https://doi.org/10.1007/s10115-017-1049-x

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10115-017-1049-x

Keywords

Navigation