Skip to main content

Isolating Stock Prices Variation with Neural Networks

  • Conference paper
  • 1371 Accesses

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 43))

Abstract

In this study we aim to define a mapping function that relates the general index value among a set of shares to the prices of individual shares. In more general terms this is problem of defining the relationship between multivariate data distributions and a specific source of variation within these distributions where the source of variation in question represents a quantity of interest related to a particular problem domain. In this respect we aim to learn a complex mapping function that can be used for mapping different values of the quantity of interest to typical novel samples of the distribution. In our investigation we compare the performance of standard neural network based methods like Multilayer Perceptrons (MLPs) and Radial Basis Functions (RBFs) as well as Mixture Density Networks (MDNs) and a latent variable method, the General Topographic Mapping (GTM). According to the results, MLPs and RBFs outperform MDNs and the GTM for this one-to-many mapping problem.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Rumelhart, D.E., Hinton, D.E., Williams, R.J.: Learning representations by back-propagation errors. Nature 323, 533–536 (1986)

    Article  Google Scholar 

  2. Powell, M.J.D.: Radial basis functions for multivariable interpolation: A review. In: IMA Conference on Algorithms for the approximation of Functions and Data, pp. 143–167. RMCS, Shrivenham (1985)

    Google Scholar 

  3. Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, New York (1995)

    MATH  Google Scholar 

  4. Bishop, C.M.: Mixture Density Networks. Technical Report NCRG/94/004, Neural Computing Research Group, Aston University (1994), http://research.microsoft.com/~cmbishop/downloads/Bishop-NCRG-94-004.ps

  5. Bishop, C.M., Svensén, M., Williams, C.K.I.: GTM: The Generative Topographic Mapping. Neural Computation 10(1), 215–234 (1998)

    Article  MATH  Google Scholar 

  6. Richmond, K.: Mixture Density Networks, Human articulatory data and acoustic-to-articulatory inversion of continuous speech (2001), http://www.cstr.ed.ac.uk/downloads/publications/2001/Richmond_2001_a.ps

  7. Bartholomew, D.J.: Latent Variable Models and Factor Analysis. Charles Griffin & Company Ltd., London (1987)

    MATH  Google Scholar 

  8. Dempster, A., Laird, N., Rubin, D.: Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society. Series B 39(1), 1–38 (1977)

    MathSciNet  MATH  Google Scholar 

  9. Carreira-Perpinan, M.A.: One-to-many mappings, continuity constraints and latent variable models. In: Proc. IEE Colloquium on Applied Statistical Pattern Recognition, Birmingham, pp. 14/1–14/6 (1999)

    Google Scholar 

  10. MacKay, D.J.C., Gibbs, M.N.: Density networks. In: Proceedings of Society for General Microbiology, Edinburgh (1997)

    Google Scholar 

  11. Zeng, X., Yeung, D.S.: A Quantified Sensitivity Measure for Multilayer Perceptron to Input Perturbation. Neural Computation 15, 183–212 (2003)

    Article  MATH  Google Scholar 

  12. FTSE Guide to UK Calculation Methods, http://www.ftse.com/Indices/UK_Indices/Downloads/uk_calculation.pdf

  13. Møler, M.: A scaled conjugate gradient algorithm for fast supervised learning. Neural Networks 6(4), 525–533 (1993)

    Article  Google Scholar 

  14. Carreira-Perpiñán, M.Á.: Continuous latent variable models for dimensionality reduction and sequential data reconstruction. PhD thesis, Dept. of Computer Science, University of Sheffield, UK (2001), http://faculty.ucmerced.edu/mcarreira-perpinan/papers/phd-thesis.html

  15. Brouwer, R.K.: Feed-forward neural network for one-to-many mappings using fuzzy sets. Neurocomputing 57, 345–360 (2004)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Draganova, C., Lanitis, A., Christodoulou, C. (2009). Isolating Stock Prices Variation with Neural Networks. In: Palmer-Brown, D., Draganova, C., Pimenidis, E., Mouratidis, H. (eds) Engineering Applications of Neural Networks. EANN 2009. Communications in Computer and Information Science, vol 43. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-03969-0_37

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-03969-0_37

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-03968-3

  • Online ISBN: 978-3-642-03969-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics