Skip to main content

Metric Effects Based on Fluctuations in Values of k in Nearest Neighbor Regressor

  • Conference paper
  • First Online:
Data Intelligence and Cognitive Informatics

Part of the book series: Algorithms for Intelligent Systems ((AIS))

Abstract

Regression branch of machine learning purely focuses on prediction of continuous values. The supervised learning branch has many regression-based methods with parametric and nonparametric learning models. In this paper, we aim to target a very subtle point related to distance-based regression model. The distance-based model used is k-nearest neighbors regressor which is a supervised nonparametric method. The point that we want to prove is the effect of k parameter of the model and its fluctuations affecting the metrics. The metrics that we use are root mean squared error and R-squared goodness of fit with their visual representation of values with respect to k values.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 259.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 329.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 329.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Cormack RM (1971) A review of classification. J Roy Stat Soc Ser A (General) 134(3):321–367. http://www.jstor.org/stable/2344237

  2. Maulud D, Abdulazeez AM (2020) A review on linear regression comprehensive in machine learning. J Appl Sci Technol Trends 1(4):140–147. https://doi.org/10.38094/jastt1457, https://jastt.org/index.php/jasttpath/article/view/57

  3. Kotsiantis SB (2007) Supervised machine learning: a review of classification techniques. In: Proceedings of the 2007 conference on emerging artificial intelligence applications in computer engineering: real word AI systems with applications in EHealth, HCI, information retrieval and pervasive technologies. IOS Press, NLD, pp 3–24

    Google Scholar 

  4. Längkvist M, Karlsson L, Loutfi A (2014) A review of unsupervised feature learning and deep learning for time-series modeling. Pattern Recogn Lett 42:11–24

    Article  Google Scholar 

  5. Bashir F, Wei HL (2015) Parametric and non-parametric methods to enhance prediction performance in the presence of missing data. In: 2015 19th international conference on system theory, control and computing (ICSTCC), pp 337–342. https://doi.org/10.1109/ICSTCC.2015.7321316

  6. Wallisch C, Bach P, Hafermann L, Klein N, Sauerbrei W, Steyerberg EW, Heinze G, Rauch G (2022) On behalf of topic group 2 of the STRATOS initiative: review of guidance papers on regression modeling in statistical series of medical journals. PLOS ONE 17(1):1–20. https://doi.org/10.1371/journal.pone.0262918

  7. Cramer JS (2002) The origins of logistic regression

    Google Scholar 

  8. Gupta A, Soni H, Joshi R, Laban RM (2022) Discriminant analysis in contrasting dimensions for polycystic ovary syndrome prognostication. arXiv preprint arXiv:2201.03029

  9. Taunk K, De S, Verma S, Swetapadma A (2019) A brief review of nearest neighbor algorithm for learning and classification. In: 2019 international conference on intelligent computing and control systems (ICCS), pp 1255–1260. https://doi.org/10.1109/ICCS45141.2019.9065747

  10. Hearst M, Dumais S, Osuna E, Platt J, Scholkopf B (1998) Support vector machines. IEEE Intell Syst Appl 13(4):18–28. https://doi.org/10.1109/5254.708428

  11. Quinlan JR (1986) Induction of decision trees. Mach Learn 1(1):81–106. https://doi.org/10.1023/A:1022643204877

  12. Kanvinde N, Gupta A, Joshi R (2022) Binary classification for high dimensional data using supervised non-parametric ensemble method. arXiv preprint arXiv:2202.07779

  13. Gupta AM, Shetty SS, Joshi RM, Laban RM (2021) Succinct differentiation of disparate boosting ensemble learning methods for prognostication of polycystic ovary syndrome diagnosis. In: 2021 international conference on advances in computing, communication, and control (ICAC3). IEEE, pp 1–5

    Google Scholar 

  14. Tibshirani R (1996) Regression shrinkage and selection via the lasso. J Roy Stat Soc Ser B (Methodological) 58(1):267–288. http://www.jstor.org/stable/2346178

  15. Hoerl AE, Kennard RW (2000). Ridge regression: biased estimation for nonorthogonal problems. Technometrics 42(1):80–86. http://www.jstor.org/stable/1271436

  16. Sheather SJ (2004) Density estimation. Stat Sci 19(4):588–597. http://www.jstor.org/stable/4144429

  17. Liberti L, Lavor C, Maculan N, Mucherino A (2014) Euclidean distance geometry and applications. SIAM Rev 56:3–69

    Article  MathSciNet  MATH  Google Scholar 

  18. Ranjitkar HS, Karki S (2016) Comparison of A*, Euclidean and Manhattan distance using influence map in MS. Pac-Man

    Google Scholar 

  19. Norouzi M, Fleet DJ, Salakhutdinov RR (2012) Hamming distance metric learning. In: Pereira F, Burges CJC, Bottou L, Weinberger KQ (eds) Advances in neural information processing systems, vol 25. Curran Associates, Inc., USA

    Google Scholar 

  20. Sammut C, Webb GI (2011) Encyclopedia of machine learning. Springer Science & Business Media, Berlin

    Google Scholar 

  21. Chai T, Draxler RR (2014) Root mean square error (RMSE) or mean absolute error (MAE)? Arguments against avoiding RMSE in the literature. Geoscientific Model Dev 7(3):1247–1250. https://doi.org/10.5194/gmd-7-1247-2014

  22. Colin Cameron A, Windmeijer FA (1997) An R-squared measure of goodness of fit for some common nonlinear regression models. J Econometrics 77(2):329–342

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Abhishek Gupta .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Gupta, A., Joshi, R., Kanvinde, N., Gerela, P., Laban, R.M. (2023). Metric Effects Based on Fluctuations in Values of k in Nearest Neighbor Regressor. In: Jacob, I.J., Kolandapalayam Shanmugam, S., Izonin, I. (eds) Data Intelligence and Cognitive Informatics. Algorithms for Intelligent Systems. Springer, Singapore. https://doi.org/10.1007/978-981-19-6004-8_12

Download citation

Publish with us

Policies and ethics