Skip to main content

Investigating Long Short-Term Memory Networks for Various Pattern Recognition Problems

  • Conference paper
Machine Learning and Data Mining in Pattern Recognition (MLDM 2014)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 8556))

Abstract

The purpose of this paper is to further investigate how and why long short-term memory networks (LSTM) perform so well on several pattern recognition problems. Our contribution is three-fold. First, we describe the main highlights of the LSTM architecture, especially when compared to standard recurrent neural networks (SRN). Second, we give an overview of previous studies to analyze the behavior of LSTMs on toy problems and some realistic data in the speech recognition domain. Third, the behavior of LSTMs is analyzed on novel problems which are relevant for pattern recognition research. Thereby, we analyze the ability of LSTMs to classify long sequences containing specific patterns at an arbitrary position on iteratively increasing the complexity of the problem under constant training conditions. We also compare the behavior of LSTMs to SRNs for text vs. non-text sequence classification on a real-world problem with significant non-local time-dependencies where the features are computed only locally. Finally, we discuss why LSTMs with standard training methods are not suited for the task of signature verification.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Eck, D., Schmidhuber, J.: Finding temporal structure in music: Blues improvisation with lstm recurrent networks. In: Proceedings of the 2002 IEEE Workshop on Neural Networks for Signal Processing XII, pp. 747–756. IEEE (2002)

    Google Scholar 

  2. Gers, F.A., Schmidhuber, J., Cummins, F.: Learning to forget: Continual prediction with LSTM. Neural Computation 12, 2451–2471 (1999)

    Article  Google Scholar 

  3. Gers, F.A., Schmidhuber, J.: LSTM recurrent networks learn simple context free and context sensitive languages. IEEE Transactions on Neural Networks 12, 1333–1340 (2001)

    Article  Google Scholar 

  4. Gers, F.A., Schraudolph, N.N., Schmidhuber, J.: Learning precise timing with LSTM recurrent networks. Journal of Machine Learning Research 3, 115–143 (2002)

    MathSciNet  Google Scholar 

  5. Graves, A., Liwicki, M., Fernandez, S., Bertolami, R., Bunke, H., Schmidhuber, J.: A novel connectionist system for unconstrained handwriting recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 31(5), 855–868 (2009)

    Article  Google Scholar 

  6. Graves, A.: A comparison of network architectures. In: Graves, A. (ed.) Supervised Sequence Labelling with Recurrent Neural Networks. SCI, vol. 385, pp. 49–58. Springer, Heidelberg (2012)

    Chapter  Google Scholar 

  7. Graves, A., Bruegge, B., Schmidhuber, J., Kramer, S.: Supervised Sequence Labelling with Recurrent Neural Networks. Ph.D. thesis, Technische Universitaet Muenchen (2008)

    Google Scholar 

  8. Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Networks: The Official Journal of the International Neural Network Society 18(5-6), 602–610 (2005)

    Article  Google Scholar 

  9. Hochreiter, S., Heusel, M., Obermayer, K.: Fast model-based protein homology detection without alignment. Bioinformatics 23(14), 1728–1736 (2007)

    Article  Google Scholar 

  10. Hochreiter, S., Schmidhuber, J.: Long Short-Term Memory. Neural Comput 9(8), 1735–1780 (1997)

    Article  Google Scholar 

  11. Indermuehle, E., Liwicki, M., Bunke, H.: IAMonDo-database: an online handwritten document database with non-uniform contents. In: 9th Int. Workshop on Document Analysis Systems, pp. 97–104 (2010)

    Google Scholar 

  12. Liwicki, M., Malik, M.I., Heuvel, C., Chen, X., Berger, C., Stoel, R., Blumenstein, M., Found, B.: Signature Verification Competition for Online and Offline Skilled Forgeries (SigComp 2011). In: 2011 International Conference on Document Analysis and Recognition (ICDAR), pp. 1480–1484. IEEE (2011)

    Google Scholar 

  13. Marti, U.V., Bunke, H.: Using a statistical language model to improve the performance of an HMM-based cursive handwriting recognition system. Int. Journal of Pattern Recognition and Artificial Intelligence 15, 65–90 (2001)

    Article  Google Scholar 

  14. Otte, S., Liwicki, M., Krechel, D., Dengel, A.: Local feature based online mode detection with recurrent neural networks. In: 2012 International Conference on Frontiers in Handwriting Recognition, ICFHR (2012)

    Google Scholar 

  15. Robinson, A.J., Fallside, F.: The utility driven dynamic error propagation network. Tech. Rep. CUED/F-INFENG/TR.1, Cambridge University Engineering Department, Cambridge (1987)

    Google Scholar 

  16. Schmidhuber, J., Gers, F., Eck, D.: Learning nonregular languages: A comparison of simple recurrent networks and lstm. Neural Computation 14 (2002)

    Google Scholar 

  17. Schmidhuber, J., Wierstra, D., Gagliolo, M., Gomez, F.: Training recurrent networks by evolino. Neural Comput. 19(3), 757–779 (2007)

    Article  MATH  Google Scholar 

  18. Schraudolph, N.N.: A fast, compact approximation of the exponential function. Neural Computation 11, 4–11 (1998)

    Google Scholar 

  19. Tiin, C., Omlin, C.: LSTM recurrent neural networks for signature veri cation. In: Proc. Southern African Telecommunication Networks & Applications Conference (2003)

    Google Scholar 

  20. Tiin, C.: LSTM Recurrent Neural Networks for Signature Veri cation: A Novel Approach. LAP LAMBERT Academic Publishing (2012)

    Google Scholar 

  21. Williams, R.J., Zipser, D.: Gradient-Based learning algorithms for recurrent networks and their computational complexity (1995)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Otte, S., Liwicki, M., Krechel, D. (2014). Investigating Long Short-Term Memory Networks for Various Pattern Recognition Problems. In: Perner, P. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2014. Lecture Notes in Computer Science(), vol 8556. Springer, Cham. https://doi.org/10.1007/978-3-319-08979-9_37

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-08979-9_37

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-08978-2

  • Online ISBN: 978-3-319-08979-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics