Skip to main content

Abstract

This introductory chapter reviews the basics of artificial-neural-network theory, discusses various aspects of the hardware implementation of neural networks (in both ASIC and FPGA technologies, with a focus on special features of artificial neural networks), and concludes with a brief note on performance-evaluation. Special points are the exploitation of the parallelism inherent in neural networks and the appropriate implementation of arithmetic functions, especially the sigmoid function. With respect to the sigmoid function, the chapter includes a significant contribution.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. U. Ruckert. 2002. ULSI architectures for artificial neural networks. IEEE Micro (May–June): 10–19.

    Google Scholar 

  2. J. Buddefeld and K. E. Grosspietsch. 2002. Intelligent-memory architectures for artificial neural networks. IEEE Micro (May–June): 32–40.

    Google Scholar 

  3. G. Danese, F. Leoporati, and S. Ramat. 2002. A parallel neural processor for real-time applications. IEEE Micro (May–June): 20–31.

    Google Scholar 

  4. A. R. Omondi. Computer Arithmetic Systems: Algorithms, Architecture, and Implementations. Prentice-Hall, UK, 1994.

    Google Scholar 

  5. T. Nordstrom and B. Svensson. 1991. Using and designing massively parallel computers for artificial neural networks. Journal of Parallel and Distributed Computing, 14:260–285.

    Article  Google Scholar 

  6. Y. Hirai. 1993. Hardware implementations of neural networks in Japan. Neurocomputing, 5:3–16.

    Article  Google Scholar 

  7. N. Sundarajan and P. Satchandran. 1998. Parallel Architectures for Artificial Neural Networks. IEE Press, California.

    Google Scholar 

  8. D. Hammerstom. 1991. A highly parallel digital architecture for neural network simulation. In: J.D. Delgado-Frias and W.R. Moore, Eds., VLSI for Artificial Intelligence and Neural Networks, Plenum Press.

    Google Scholar 

  9. R. M. Corless, G. H. Gonnet, D. E. G. Hare, and D. J. Jeffrey, and D. E. Knuth. 1996. On the Lambert W Function. Advances in Computational Mathematics, 12:329–359.

    Article  MathSciNet  Google Scholar 

  10. M. Ito, N. Takagi, and S. Yajima. 1997. Efficient initial approximation for multiplicative division and square-root by a multiplication with operand modification. IEEE Transactions on Computers, 46(4):95–498.

    Article  MathSciNet  Google Scholar 

  11. J. M. Muller. 1997. Elementary Functions: Algorithms and Implementation. Birkhauser, Boston, USA.

    Google Scholar 

  12. S. M. Pizer and V. L. Wallace. 1983. To Compute Numerically. Little, Brown, and Co., Boston, USA.

    Google Scholar 

  13. M. Bajger and A. R. Omondi. 2005. Low-cost, high-speed implementations of square-root, exponential and sigmoidal function-evaluations. Submitted for publication.

    Google Scholar 

  14. S. Vassiliadis, M. Zhang, and J. G. Delgado-Frias. 2000. Elementary function generators for neural network emulators. IEEE Transactions on Neural Networks, 11(6):1438–1449.

    Article  Google Scholar 

  15. K. Basterretxea, J. M. Tarela, and I. del Campo. 2004. Approximation of sigmoid function and the derivative for hardware implementation of artificial neurons. IEEE Proceedings — Circuits, Devices, and Systems, 151(1):18–24.

    Article  Google Scholar 

  16. O. Mencer and W. Luk. 2004. Parameterized high throughput function evaluation for FPGAs. Journal of VLSI Signal Processing, 36:17–25.

    Article  Google Scholar 

  17. J. L. Holt and J. N. Hwang. 1993. Finite-precision error analysis of neural network hardware implementations. IEEE IEEE Transactions on Computers, 42(3):280–290.

    Google Scholar 

  18. A. R. Omondi. 2000. Neurocomputers: a dead end? International Journal of Neural Systems, 10(6):475–481.

    Google Scholar 

  19. J. L. Hennessy and D. A. Patterson. 2002. Computer Architecture: A Quantitative Approach. Morgan Kaufmann.

    Google Scholar 

  20. SPEC. Standard Performance Evaluation Corporation. (www.spec.org )

    Google Scholar 

  21. Xilinx. 2004. Virtex-4 User Guide.

    Google Scholar 

  22. Xilinx. 2004. XtremeDSP Design Considerations: User Guide.

    Google Scholar 

  23. A. P. Preethy, D. Radhakrishnan, A. R. Omondi. Mar 2001. A high-performance residue-number-system multiply-accumulate unit. In: 11th ACM Great Lakes Symposium on VLSI (Purdue, Indiana, USA), pp 145–149.

    Google Scholar 

  24. Waterloo Maple Inc. Maple 8 Programming Guide, 2002.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer

About this chapter

Cite this chapter

Omondi, A.R., Rajapakse, J.C., Bajger, M. (2006). FPGA Neurocomputers. In: Omondi, A.R., Rajapakse, J.C. (eds) FPGA Implementations of Neural Networks. Springer, Boston, MA . https://doi.org/10.1007/0-387-28487-7_1

Download citation

  • DOI: https://doi.org/10.1007/0-387-28487-7_1

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-0-387-28485-9

  • Online ISBN: 978-0-387-28487-3

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics