Dimensionality Reduction Technique to Solve E-Crime Motives

  • R. AartheeEmail author
  • D. Ezhilmaran
Conference paper
Part of the Trends in Mathematics book series (TM)


The dimensionality reduction technique is a great way of math or statistics to minimize the size of data as much as possible, as little information is possible. With a large number of variables, the dispersed matrix may be too large to be studied and interpreted correctly. There will be too much correlation between the variables to be considered. Graphics data is also not particularly useful because the dataset is large. To interpret the more meaningful data, it is essential to reduce the number of variables to a few linear combinations. Each linear combination will correspond to one major component. Dimensionality reduction technique used to transform dataset onto a lower dimensional subspace for visualization and exploration. This technique is also called as principal component analysis. In this article, we are developing an analysis of the essential elements of the cybercrime motives database and discovering some of the high reasons for increasing cybercriminals.


  1. 1.
    Saporta,G., Niang,N.,: Principal component analysis: application to statistical process control. Data analysis, 1–23(2009)Google Scholar
  2. 2.
    O’Rourke,N., Psych,R., Hatcher,L.,: A step-by-step approach to using SAS for factor analysis and structural equation modeling. Sas Institute,(2013).Google Scholar
  3. 3.
    Dunteman,G.H., Principal components analysis. Sage,69,(1989)Google Scholar
  4. 4.
    Jolliffe,I.T., Principal component analysis and factor analysis. In Principal component analysis.Springer, New York, NY. 115–128(1986)Google Scholar
  5. 5.
    Shlens,J., A tutorial on principal component analysis. arXiv preprint, 1404–1100 (2014)Google Scholar
  6. 6.
    Abdi,H., Williams,L.J., Principal component analysis. Wiley interdisciplinary reviews: computational statistics, 2(4). 433–459 (2010)CrossRefGoogle Scholar
  7. 7.
    Olive,D.J., Robust multivariate analysis. Springer.(2018)Google Scholar
  8. 8.
    Elhamifar,E., Vidal,R.,Sparse subspace clustering. In Computer Vision and Pattern Recognition, 2009. CVPR 2009, 2790–2797(2009)Google Scholar
  9. 9.
    Liu,W., Zhang,H., Tao,D., Wang,Y., Lu.K., Large-scale paralleled sparse principal component analysis. Multimedia Tools and Applications, 75(3). 1481–1493 (2016)CrossRefGoogle Scholar
  10. 10.
    Guan,N., Tao,D., Luo,Z., Yuan,B.,Online nonnegative matrix factorization with robust stochastic approximation. IEEE Transactions on Neural Networks and Learning Systems, 23(7). 1087–1099(2012)CrossRefGoogle Scholar
  11. 11.
    Xu,C., Tao,D., Large-margin multi-view information bottleneck. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(8).1559–1572(2014)CrossRefGoogle Scholar
  12. 12.
    Tao,D., Li,X., Wu,X., Maybank,S.J., General tensor discriminant analysis and gabor features for gait recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(10)(2007)CrossRefGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2018

Authors and Affiliations

  1. 1.VITVelloreIndia

Personalised recommendations