Abstract
The use of principal component analysis in preprocessing neural network input data is explored. Four preprocessing schemes are compared in an example problem, and the theoretical basis for the results are discussed. A preconditioning method for the principal components is introduced here, combining normalisation and improved conditioning. The techniques are applied to an object location problem in diffraction tomography. The spectral analysed scattered field from an irradiated object form the input to a Multilayer Perceptron neural network, trained by backpropagation to calculate the coordinates of the object's centre in 2D.
Similar content being viewed by others
References
Jacobs RA. Increased rates of convergence through learning rate adaption. Neural Networks 1988; 1: 295–307
Rigler AK, Irvine JM, Vogl, TP. Rescaling of variables in back propagation learning. Neural Networks 1991; 4: 225–229
Morrison DF. Multivariate Statistical Methods, 2nd ed., McGraw-Hill, New York, NY, 1976
Gonzalez RC, Wintz P. Digital Image Processing, 2nd ed., Addison-Wesley, Reading MA, 1987
Rumelhart DE, Hinton GE, Williams RJ. Nature 1986; 323: 533–536
Oja E. A simplified neuron model as a principal component analyzer. J Math Biol 1982; 15: 267–273
Oja E. Neural networks, principal components, and subspaces. Int J Neural Syst 1989; 1: 61–68
Bourlard H, Kamp Y. Auto-association by multilayer perceptrons and singular value decomposition, Biol Cybern 1988; 59: 291–294
Baldi P, Hornik K. Neural networks and principal component analysis: learning from examples without local minima. Neural Networks 1989; 2: 53–58
Young TY, Fu KS, eds. Handbook of pattern recognition and image processing, Academic Press, New York, NY, 1986
Farhat NH. Microwave diversity imaging and automated target identification based on models of neural networks. Proc IEEE 1989; 77: 670–681
Roth MW. Survey of neural network technology for automatic target recognition. IEEE Trans Neural Networks 1990; 1: 28–43
Press WH, Flannery BP, Teukolsky SA, Vetterling WT. Numerical recipes. Cambridge University Press, Cambridge, 1987
Haykin S. Adaptive filter theory, 2nd ed., Prentice-Hall, Englewood Cliffs, NJ, 1991
Luenberger DG. Introduction to linear & nonlinear programming, Addison-Wesley, Reading, MA, 1973
van der Sluis A, van der Vorst HA, The rate of convergence of conjugate gradients. Numer Math 1986; 48: 543–560
Malki HA, Moghaddamjoo A. Using the Karhunen-Loève transformation in the backpropagation training algorithm. IEEE Trans Neural Networks 1991; 2: 162–165
Pan SX, Kak AC. A computational study of reconstruction algorithms for diffraction tomography. IEEE Trans Acoust Speech & Signal Process 1983; 31: 1262–1275
Wasserman PD. Neural computing, Van Nostrand Reinhold, New York, NY, 1989
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Azoff, E.M. Neural network principal components preprocessing and diffraction tomography. Neural Comput & Applic 1, 107–114 (1993). https://doi.org/10.1007/BF01414430
Received:
Issue Date:
DOI: https://doi.org/10.1007/BF01414430