Optimization-Based Effective Feature Set Selection in Big Data

Conference paper
Part of the Lecture Notes in Networks and Systems book series (LNNS, volume 118)


Of late, the data mining has appeared on the arena as an ideal form of knowledge discovery crucial for the purpose of providing appropriate solutions to an assortment of issues in a specified sphere. In this regard, the classification represents an effective method deployed with a view to locating several categories of anonymous data. Further, the feature selection has significantly showcased its supreme efficiency in a host of applications by effectively ushering in easier and more all-inclusive remodel, augmenting the learning performance, and organizing fresh and comprehensible data. However, of late, certain severe stumbling blocks have cropped up in the arena of feature selection, in the form of certain distinctive traits of significant of big data, like the data velocity and data variety. In the document, a sincere effort is made to successfully address the prospective problems encountered by the feature selection in respect of big data analytics. Various tests conducted have upheld the fact that the oppositional grasshopper techniques are endowed with the acumen of effectively extracting the requisite features so as to achieve the preferred outcome Further, enthusing experimental outcomes have revealed the fact only a trivial number of hidden neurons are necessary for the purpose of the feature selection to effectively appraise the quality of an individual, which represents a chosen subset of features.


Classification Optimization Oppositional grasshopper 


  1. 1.
    Wu X, Zhu X, Wu G-Q, Ding W (2014) Data mining with big data. IEEE Trans Knowl Data Eng 26(1):97–107CrossRefGoogle Scholar
  2. 2.
    Slavakis K, Giannakis GB, Mateos G (2014) Modeling and optimization for big data analytics: (statistical) learning tools for our era of data deluge. IEEE Signal Process Mag 31(5):18–31CrossRefGoogle Scholar
  3. 3.
    Anjaria M, Guddeti RMR (2014) Influence factor based opinion mining of Twitter data using supervised learning. In process of Sixth International Conference on Communication Systems and Networks (COMSNETS), 2014, pp 1–8Google Scholar
  4. 4.
    Jiang Xudong (2011) Linear subspace learning-based dimensionality reduction. IEEE Signal Process Mag 28(2):16–26CrossRefGoogle Scholar
  5. 5.
    ur Rehman MH, Chang V, Batool A, Wah TY (2016) Big data reduction framework for value creation in sustainable enterprises. Int J Inf Manag 36(6):917–928CrossRefGoogle Scholar
  6. 6.
    Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E (2015) Deep learning applications and challenges in big data analytics. J Big Data 2(1):1CrossRefGoogle Scholar
  7. 7.
    Zou Q, Zeng J, Cao L, Ji R (2016) A novel features ranking metric with application to scalable visual and bioinformatics data classification. Neurocomputing 173:346–354CrossRefGoogle Scholar
  8. 8.
    Diamantoulakis PD, Kapinas VM, Karagiannidis GK (2015) Big data analytics for dynamic energy management in smart grids. Big Data Res 2(3):94–101CrossRefGoogle Scholar
  9. 9.
    Phinyomark A, Phukpattaranont P, Limsakul C (2012) Feature reduction and selection for EMG signal classification. Exp Syst Appl 39(8):7420–7431CrossRefGoogle Scholar
  10. 10.
    Chen CLP, Zhang C-Y (2014) Data-intensive applications, challenges, techniques and technologies: a survey on Big Data. Inf Sci 275:314–347CrossRefGoogle Scholar

Copyright information

© Springer Nature Singapore Pte Ltd. 2020

Authors and Affiliations

  1. 1.Research Scholar, School of Computing ScienceVels Institute of Science, Technology and Advanced Studies, (VISTAS)ChennaiIndia
  2. 2.Associate Professor, Department of Computer Science and ApplicationsSRM Institute for Training and DevelopmentChennaiIndia
  3. 3.Assistant Professor, School of Computing ScienceVels Institute of Science, Technology & Advanced Studies, (VISTAS)ChennaiIndia

Personalised recommendations