Skip to main content

Part of the book series: Mathematics and Its Applications ((MAIA,volume 574))

  • 573 Accesses

Abstract

In recent years there has been an enormous development of the applications of neural networks for solving practical problems. Neural networks take their name and structure from the neuronal system of animals and humans. They can be described as a unit with an input and corresponding output. The input is taken from any time series composed by vectors of n variables containing the results of some experiment at time t and the output might be some state variable which summarizes the conclusion of the experiment at time t + 1 represented by a vector of n variables. The transformation which associates the output to a given input is defined in analogy with the behavior of the real neurones. For example the simplest structure is with one unit of m intermediate neurones while the input is described by n neurones and the output by q neurones. The input neurones are connected by synaptic weights w 1 ij , i = 1,...,m, j = 1,...,n to the neurones of the inner layer and the neurones of the inner layer are connected to the neurones of the output layer by other synaptic weights w 2 li , l = 1,..., q, i = 1,..., m. The input vector ξ j , j = 1,..., n is transformed into a vector of outputs h i , i = 1,..., m of the intermediate level by the law defined by the weightsw 1 ij :

$$ {h_i} = f(\sum\limits_{j = 1}^n {w_{ij}^1} {\xi _j}) $$

, where f is the transfer function of the neuron which has the simple form

$$ f(x) = \frac{1}{{1 + \exp ( - \lambda x)}},x \in R $$

withλ > 0 a given constant. The output of the network is the vector γl, l = 1,...q given by

$$ \gamma l = f(\sum\limits_{i = 1}^m {w_{li}^2} {h_i}) $$

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2004 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Khrennikov, A.Y., Nilson, M. (2004). P-Adic Neural Networks. In: P-adic Deterministic and Random Dynamics. Mathematics and Its Applications, vol 574. Springer, Dordrecht. https://doi.org/10.1007/978-1-4020-2660-7_8

Download citation

  • DOI: https://doi.org/10.1007/978-1-4020-2660-7_8

  • Publisher Name: Springer, Dordrecht

  • Print ISBN: 978-90-481-6698-5

  • Online ISBN: 978-1-4020-2660-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics