Skip to main content

An Improved Backpropagation Algorithm Using Absolute Error Function

  • Conference paper
Advances in Neural Networks – ISNN 2005 (ISNN 2005)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 3496))

Included in the following conference series:

Abstract

An improved backpropagation algorithm is proposed by using the Lyapunov method to minimize the absolution error function. The improved algorithm can make both error and gradient approach zero so that the local minima problem can be avoided. In addition, since the absolute error function is used, this algorithm is more robust and faster for learning than the backpropagation with the traditional square error function when target signals include some incorrect data. This paper also proposes a method of using Lyapunov stability theory to derive a learning algorithm which directly minimize the absolute error function.

This work was supported by National Science Foundation of China under Grant 60471055 and Specialized Research Fund for the Doctoral Program of Higher Education under Grant 20040614017.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Werbos, P.J.: The Roots of Backpropagation. Wiley, New York (1994)

    Google Scholar 

  2. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning Internal Representations by Error Propagation. In: Parallel distributed Processing, pp. 318–362. MIT Press, Cambridge (1986)

    Google Scholar 

  3. Taji, K., Miyake, T., Tarmura, H.: On Error Backpropagation Algorithm Using Absolute Error Function. In: IEEE SMC 1999 Conference Proceedings, vol. 5, pp. 12–15 (1999)

    Google Scholar 

  4. Wang, X.G., Tang, Z., Tamura, H., Ishii, M.: A Modified Error Function for the Backpropagation Algorithm. Neurocomputing 57, 477–484 (2004)

    Article  Google Scholar 

  5. Wang, X.G., Tang, Z., Tamura, H., Ishii, M., Sum, W.D.: An Improved Backpropagation Algorithm to Avoid the Local Minima Problem. Neurocomputing 56, 455–460 (2004)

    Article  Google Scholar 

  6. Owen, C.B., Abunawass, A.M.: Application of Simulated Annealing to the Backpropagation Model Improves Convergence. In: Proceeding of the SPIE Conference on the Science of Artificial Neural Networks, vol. II, pp. 269–276 (1993)

    Google Scholar 

  7. Von Lehmen, A., Paek, E.G., Liao, P.F., Marrakchi, A., Patel, J.S.: Factors Influencing Learning by Backpropagation. In: Proceedings of the IEEE International Conference On Neural Networks, vol. I, pp. 335–341 (1988)

    Google Scholar 

  8. Fukuoka, Y., Matsuki, H., Minamitani, H., Ishida, A.: A Modified Backpropagation Method to Avoid False Local Minima. Neural Networks 11, 1059–1072 (1998)

    Article  Google Scholar 

  9. Vitela, J.E., Reifman, J.: Premature Saturation in Backpropagation Networks: Mechanism and Necessary Conditions. Neural Networks 10, 721–735 (1997)

    Article  Google Scholar 

  10. Wand, C., Principe, J.C.: Training Neural Networks with Additive Noise in the Desired Signal. IEEE Trans. Neural Networks 10, 1511–1517 (1999)

    Article  Google Scholar 

  11. Battiti, R., Masulli, F.: BFGS Optimization for Faster and Automated Supervised Learning. In: Proceedings of the Internatioanl Neural Network Conference, pp. 757–760. Kluwer Academic Publishers, Paris (1990)

    Google Scholar 

  12. Kollias, S., Anastassiou, D.: An Adaptive Least Squares Algorithm for Efficient Training of Artificial Neural Networks. IEEE Trans. on Circuits and systems 36, 1092–1101 (1989)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Lv, J., Yi, Z. (2005). An Improved Backpropagation Algorithm Using Absolute Error Function. In: Wang, J., Liao, X., Yi, Z. (eds) Advances in Neural Networks – ISNN 2005. ISNN 2005. Lecture Notes in Computer Science, vol 3496. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11427391_93

Download citation

  • DOI: https://doi.org/10.1007/11427391_93

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-25912-1

  • Online ISBN: 978-3-540-32065-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics