Abstract
In the past few years, the problem of processing of big data in less time and with great accuracy has been a major concern in the world of machine learning. Problems like overfitting and underfitting have also been on the troubling side of projects in the years leading to the current stages of artificial intelligence. This model works on the principle of dimensionality reduction, a potential solution to the above problems, with the growth of artificial intelligence and machine learning, data has become one of the most important assets and yet the biggest challenge. Over the years fast and accurate processing of data has been a constant pursuit, and this is where dimensionality reduction comes into play. In dimensionality reduction, authors reduce the dimension of the data keeping the essence and all the important points of the data intact, this leads to faster and equally accurate result. In my project, I have shown the concept of dimensionality reduction on images using the fonts dataset. Variational autoencoders are used to reduce dimension of the images from larger dimensions to smaller dimensions. The model was then checked for the training loss and the validation loss. Also, GPU based approach is used for the upscaling of the model and obtaining competent and satisfying results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Pu Y, Gan z, Henao R, Yuan X, Li C, Stevens A, Carin L (2016) Variational autoencoder for deep learning of images, labels and captions. http://papers.nips.cc/paper/6528-variational-autoencoder-for-deep-learning-of-images-labels-and-captions
Sam T, Roweis, Saul LK (2000) Nonlinear dimensionality reduction by locally linear embedding. Science 290(5500):2323–2326
Tenenbaum JB, Silva VD, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500):2319–2323
Maaten LVD, Postma E, Herik JVD (2009) Experiment dimensionality reduction: a comparative review. https://members.loria.fr/moberger/Enseignement/AVR/Exposes/TR_Dimensiereductie.pdf
Rabbani M (2002) JPEG2000: Image compression fundamentals. Standards and practice. J Electron Imaging 11(2). https://doi.org/10.1117/1.1469618
Theis L, Shi W, Cunningham A, Huszár F (2017) Lossy image compression with compressive autoencoders. https://arxiv.org/abs/1703.00395
Verma, Kumar V, Tiwari PK (2015) Removal of obstacles in Devanagari script for efficient optical character recognition. In: 2015 International conference on computational intelligence and communication networks (CICN). IEEE
Variational Autoencoders—Jordan J. https://www.jeremyjordan.me/variational-autoencoders
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Dahiya, P., Garg, S. (2022). Dimensionality Reduction Using Variational Autoencoders. In: Nanda, P., Verma, V.K., Srivastava, S., Gupta, R.K., Mazumdar, A.P. (eds) Data Engineering for Smart Systems. Lecture Notes in Networks and Systems, vol 238. Springer, Singapore. https://doi.org/10.1007/978-981-16-2641-8_24
Download citation
DOI: https://doi.org/10.1007/978-981-16-2641-8_24
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-2640-1
Online ISBN: 978-981-16-2641-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)