Skip to main content

Advertisement

Log in

Learning from Few Samples with Memory Network

  • Published:
Cognitive Computation Aims and scope Submit manuscript

Abstract

Neural networks (NN) have achieved great successes in pattern recognition and machine learning. However, the success of a NN usually relies on the provision of a sufficiently large number of data samples as training data. When fed with a limited data set, a NN’s performance may be degraded significantly. In this paper, a novel NN structure is proposed called a memory network. It is inspired by the cognitive mechanism of human beings, which can learn effectively, even from limited data. Taking advantage of the memory from previous samples, the new model achieves a remarkable improvement in performance when trained using limited data. The memory network is demonstrated here using the multi-layer perceptron (MLP) as a base model. However, it would be straightforward to extend the idea to other neural networks, e.g., convolutional neural networks (CNN). In this paper, the memory network structure is detailed, the training algorithm is presented, and a series of experiments are conducted to validate the proposed framework. Experimental results show that the proposed model outperforms traditional MLP-based models as well as other competitive algorithms in response to two real benchmark data sets.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

References

  1. Ruck DW, Rogers SK, Kabrisky M, Oxley ME, Suter BW. The multilayer perceptron as an approximation to a bayes optimal discriminant function. IEEE Trans Neural Netw 1990;1(4):296–298.

    Article  CAS  PubMed  Google Scholar 

  2. Zhang S, Huang K. Learning from few samples with memory network. International Conference on Neural Information Processing. Springer; 2016. p. 606–614.

  3. Gao F, Zhang Y, Wang J, Sun J, Yang E, Hussain A. Visual attention model based vehicle target detection in synthetic aperture radar images: a novel approach. Cogn Comput 2015;7(4):434–444.

    Article  Google Scholar 

  4. Girshick R. Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision; 2015. p. 1440–1448.

  5. Iandola FN, Han S, Moskewicz M, Ashraf K, Dally WJ, Keutzer K. 2016. Squeezenet: alexnet-level accuracy with 50x fewer parameters and 0.5 mb model size. arXiv:1602.07360.

  6. Lawrence S, Giles CL, Tsoi AC, Back AD. Face recognition: a convolutional neural-network approach. IEEE Trans Neural Netw 1997;8(1):98–113.

    Article  CAS  PubMed  Google Scholar 

  7. Lyu C, Huang K, Liang H-N. A unified gradient regularization family for adversarial examples. 2015 IEEE International Conference on Data Mining (ICDM). IEEE; 2015. p. 301–309.

  8. Oquab M, Bottou L, Laptev I, Sivic J. Learning and transferring mid-level image representations using convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2014. p. 1717–1724.

  9. Sun Y, Liang D, Wang X, Tang X. 2015. Deepid3: face recognition with very deep neural networks. arXiv:1502.00873.

  10. Rumelhart DE, Hinton GE, Williams RJ. Learning representations by back-propagating errors. Cogn Model 1988;5(3):1.

    Google Scholar 

  11. Cambria E, Hussain A, Vol. 1. Sentic computing: a common-sense-based framework for concept-level sentiment analysis. Berlin: Springer; 2015.

    Book  Google Scholar 

  12. Chu H, Huang K, Zhang R, Hussian A. Sdrnf: generating scalable and discriminative random nonlinear features from data. Big Data Anal 2016;1(1):10.

    Article  Google Scholar 

  13. Malik Z, Hussain A, Wu J. 2015. Extracting online information from dual and multi-data streams (in press). Neural Computation and Applications.

  14. Huang K, Yang H, King I, Lyu MR. Machine learning: modeling data locally and globally. Berlin: Springer; 2008.

    Google Scholar 

  15. Huang K, Yang H, King I, Lyu MR. Local learning vs. global learning: an introduction to maxi-min margin machine. Support Vector Machines: Theory and Applications. Springer; 2005. p. 113–131.

  16. Yang H, Huang K, King I, Lyu MR. Maximum margin semi-supervised learning with irrelevant data. Neural Netw 2015;70:90–102.

    Article  PubMed  Google Scholar 

  17. Wang H, Yeung D-Y. 2016. Towards bayesian deep learning: a survey. arXiv:1604.01662.

  18. Tao D, Li X, Wu X, Maybank SJ. Geometric mean for subspace selection. IEEE Trans Pattern Anal Mach Intell 2009;31(2):260–274.

    Article  PubMed  Google Scholar 

  19. Liu W, Zha ZJ, Wang Y, Lu K, Tao D. p-laplacian regularized sparse coding for human activity recognition. IEEE Trans Ind Electron 2016;63(8):5120–5129.

    Google Scholar 

  20. Gourier N, Hall D, Crowley J. Estimating face orientation from robust detection of salient facial features. International conference on pattern recognition (ICPR); 2004.

  21. Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE 1998;86(11):2278–2324.

    Article  Google Scholar 

  22. Zhang X-Y, Huang K, Liu C-L. Pattern field classification with style normalized transformation. International joint conference on artificial intelligence (IJCAI); 2011. p. 1621– 1626.

  23. Jolliffe I. Principal component analysis. Oxford: Wiley Online Library; 2002.

    Google Scholar 

  24. Tenenbaum JB, Freeman WT. Separating style and content with bilinear models. Neural Comput 2000;12(6):1247–1283.

    Article  CAS  PubMed  Google Scholar 

  25. Sarkar P, Nagy G. Style consistent classification of isogenous patterns. IEEE Trans Pattern Anal Mach Intell Jan 2005;27(1):88–98.

    Article  PubMed  Google Scholar 

  26. Grother PJ. Nist special database 19 handprinted forms and characters database. Maryland: National Institute of Standards and Technology; 1995.

    Google Scholar 

Download references

Acknowledgements

The paper was supported by National Science Foundation of China (NSFC 61473236), and Jiangsu University Natural Science Research Programme (14KJB520037).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kaizhu Huang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants performed by any of the authors.

Informed Consent

Informed consent was obtained from all individual participants included in the study.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, S., Huang, K., Zhang, R. et al. Learning from Few Samples with Memory Network. Cogn Comput 10, 15–22 (2018). https://doi.org/10.1007/s12559-017-9507-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12559-017-9507-z

Keywords

Navigation