Skip to main content

Dimensionality Reduction Using PCA and SVD in Big Data: A Comparative Case Study

  • Conference paper
  • First Online:
Future Internet Technologies and Trends (ICFITT 2017)

Abstract

With the advancement in technology, data produced from different sources such as Internet, health care, financial companies, social media, etc. are increases continuously at a rapid rate. Potential growth of this data in terms of volume, variety and velocity coined a new emerging area of research, Big Data (BD). Continuous storage, processing, monitoring (if required), real time analysis are few current challenges of BD. However, these challenges becomes more critical when data can be uncertain, inconsistent and redundant. Hence, to reduce the overall processing time dimensionality reduction (DR) is one of the efficient techniques. Therefore, keeping in view of the above, in this paper, we have used principle component analysis (PCA) and singular value decomposition (SVD) techniques to perform DR over BD. We have compared the performance of both techniques in terms of accuracy and mean square error (MSR). Comparative results shows that for numerical reasons SVD is preferred PCA. Whereas, using PCA to train the data in dimension reduction for an image gives good classification output.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Gantz, J., Reinsel, D.: IDC, The Digital Universe (2014)

    Google Scholar 

  2. Swati, A., Ade, R.: Dimensionality reduction: an effective technique for feature selection. Int. J. Comput. Appl. 117(3), 18–23 (2015)

    Google Scholar 

  3. Gupta, T.K., et al.: Dimensionality reduction techniques and its applications. J. Comput. Sci. Syst. Biol. 8(3), 170 (2015)

    Google Scholar 

  4. Person, K.: On lines and planes of closest fit to system of points in space. Philos. Mag. 2, 559–572 (1901)

    Article  Google Scholar 

  5. Hotelling, H.: Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 24(6), 417 (1933)

    Article  MATH  Google Scholar 

  6. Jollie, I.T.: Principal Component Analysis. Springer, New York (1986)

    Book  Google Scholar 

  7. Oja, E.: Simplifed neuron model as a principal component analyzer. J. Math. Biol. 15(3), 267273 (1982)

    Article  Google Scholar 

  8. Terence, D.: An optimality principle for unsupervised learning. In: NIPS, pp. 11–19 (1988)

    Google Scholar 

  9. Kung, S.Y., Diamantaras, K.I.: A neural network learning algorithm for adaptive principal component extraction (APEX). In: International Conference on Acoustics, Speech, and Signal Processing, ICASSP 1990, pp. 861–864 (1990)

    Google Scholar 

  10. Rubner, J., Tavan, P.: A self-organizing network for principal-component analysis. EPL (Europhysics Letters) 10(7), 693–696 (1989)

    Article  Google Scholar 

  11. Baldi, P., Hornik, K.: Neural networks and principal component analysis: learning from examples without local minima. Neural Netw. 2(1), 53–58 (1989)

    Article  Google Scholar 

  12. Golub, G.H., Van Loan, C.F.: Matrix Computations, 3rd edn. JHU Press, Baltimore and London (2012)

    Google Scholar 

  13. Henry, E.R., Hofrichter, J.: Singular value decomposition: application to analysis of experimental data. Methods Enzymol. 210, 129–192 (1992)

    Article  Google Scholar 

  14. Deerwester, S., Harshman, R., et al.: Indexing by latent semantic analysis. J. Am. Soc. Inf. Sci. 41(6), 391–397 (1990)

    Article  Google Scholar 

  15. Sarwar, B., et al.: Application of dimensionality reduction in recommender system-a case study. Technical report, DTIC Document (2000)

    Google Scholar 

  16. Brand, M.: Fast online SVD revisions for lightweight recommender systems. In: Proceedings of the International Conference on Data Mining, pp. 37–46. SIAM (2003)

    Google Scholar 

  17. Sarwar, B., et al.: Incremental singular value decomposition algorithms for highly scalable recommender systems. In: Fifth International Conference on Computer and Information Science, pp. 27–28 (2002)

    Google Scholar 

  18. Lichman, M.: UCI Machine Learning Repository (2013)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sudeep Tanwar .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tanwar, S., Ramani, T., Tyagi, S. (2018). Dimensionality Reduction Using PCA and SVD in Big Data: A Comparative Case Study. In: Patel, Z., Gupta, S. (eds) Future Internet Technologies and Trends. ICFITT 2017. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 220. Springer, Cham. https://doi.org/10.1007/978-3-319-73712-6_12

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-73712-6_12

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-73711-9

  • Online ISBN: 978-3-319-73712-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics