Towards Optimizing Data Analysis for Multi-dimensional Data Sets
The K-Nearest Neighbors (KNN) algorithm is a simple but powerful technique used for data analysis. It identifies existing samples in a dataset which are similar to a new sample by using a distance metric. The new sample can then be classified via a class majority voting of its most similar samples, i.e. nearest neighbors. The KNN algorithm can be applied in many fields, such as recommender systems where it can be used to group related products or predict user preferences. In many cases, the performance of the KNN algorithm diminishes as the size of the dataset increases because the number of comparisons performed increases exponentially. In this paper, we propose a KNN optimization algorithm which leverages vector space models to enhance the nearest neighbors search for a new sample. It accomplishes this enhancement by restricting the search area, and therefore reducing the number of comparisons necessary. The experimental results demonstrate significant performance improvements without compromising the algorithm’s accuracy.
KeywordsK-Nearest Neighbor KNN Vector space model Cosine similarity
- 1.Cai, Y., Ji, D., Cai, D.: A KNN research paper classification method based on shared nearest neighbor (2010)Google Scholar
- 2.Rahal, I., Perrizo, W.: An optimized approach for KNN text categorization using P-trees (2004)Google Scholar
- 3.Guo, G., Wang, H., Bell, D., Bi, Y., Greer, K.: KNN model-based approach in classification (2003)Google Scholar
- 4.Dong, T., Cheng, W., Shang, W.: The research of kNN text categorization algorithm based on eager learning (2012)Google Scholar