Artificial neural networks (ANNs), as an emerging discipline, studies or emulates the information processing capabilities of neurons of the human brain. It uses a distributed representation of the information stored in the network, and thus resulting in robustness against damage and corresponding fault tolerance (Shadbolt and Taylor, 2002). Usually, a neural network model takes an input vector X and produces output vector Y. The relationship between X and Y is determined by the network architecture. There are many forms of network architecture inspired by the neural architecture of the human brain.
In the neural network model, it is widely accepted that a three-layer back propagation neural network (BPNN) with an identity transfer function in the output unit and logistic functions in the middle-layer units can approximate any continuous function arbitrarily well given a sufficient amount of middle-layer units (White, 1990). Furthermore, in the practical applications, about 70 percent of all problems are usually trained on a three-layer back-propagation network, as revealed by Chapter 1. The backpropagation learning algorithm, designed to train a feed-forward network, is an effective learning technique used to exploit the regularities and exceptions in the training sample.
A major advantage of neural networks is their ability to provide flexible mapping between inputs and outputs. The arrangement of the simple units into a multilayer framework produces a map between inputs and outputs that is consistent with any underlying functional relationship regardless of its “true” functional form. Having a general map between the input and output vectors eliminates the need for unjustified priori restrictions that are needed in conventional statistical and econometric modeling. Therefore, a neural network is often viewed as a “universal approximator” i.e. a flexible functional form that can approximate any arbitrary function arbitrarily well, given sufficient middle-layer units and properly adjusted weights.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 2007 Springer Science+Business Media, LLC
About this chapter
Cite this chapter
(2007). Basic Learning Principles of Artificial Neural Networks. In: Foreign-Exchange-Rate Forecasting With Artificial Neural Networks. International Series in Operations Research & Management Science, vol 107. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-71720-3_2
Download citation
DOI: https://doi.org/10.1007/978-0-387-71720-3_2
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-387-71719-7
Online ISBN: 978-0-387-71720-3
eBook Packages: Business and EconomicsEconomics and Finance (R0)