skip to main content
10.1145/3055635.3056604acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlcConference Proceedingsconference-collections
research-article

An Improved k-NN Classification with Dynamic k

Published: 24 February 2017 Publication History

Abstract

In the k-NN algorithm, k is the only parameter and often set to a fixed value empirically. However, it is very difficult to choose an appropriate k in practice, and if the choice is not appropriate, the performance of k-NN will be affected greatly. In order to solve this problem, the paper proposes an improved k-NN algorithm, which is denoted as Dk-NN, by using dynamic k in replace of fixed k value. Firstly, a preprocessed step is designed and added to the traditional k-NN algorithm for determining the dynamic k interval. Then, each class's percentage of test sample is calculated within the dynamic k interval. Finally, three criterions are given to determine the class of the test sample according to the variation tendency of the percentage curves. Experimental results on real-world dataset demonstrate that the proposed algorithm is more effective than the k-NN algorithm with fixed k value.

References

[1]
Taneja, S., Gupta, C., Aggarwal, S., and Jindal, V. 2015. MFZ-KNN-A modified fuzzy based K nearest neighbor algorithm. International Conference on Cognitive Computing and Information Processing. (April, 2015), 1--5.
[2]
Dhurandhar, A., and Dobra, A. 2013. Probabilistic characterization of nearest neighbor classifier. Int. J. Mach. Learn. & Cyber. 4, 4 (August, 2013), 259--272.
[3]
Khadim, D., Fleur, M., and Gayo, D. 2016. Large scale biomedical texts classification: a knn and an esa-based approaches. J Biomed Semant. 7, 1 (December, 2016), 1--12.
[4]
Agarwal, S., and Sureka, A. 2015. Using KNN and SVM Based One-Class Classifier for Detecting Online Radicalization on Twitter. 11th International Conference, ICDCIT. (February, 2015), 431--442.
[5]
Cover, T. M., and Hart, P. E. 1967. Nearest neighbor pattern classification. IEEE Trans Inf Theory. 13,1, (1967), 21--27.
[6]
Mary-Huard, T., and Robin, S. 2009. Tailored aggregation for classification. IEEE Trans Pattern Anal Mach Intell. 31, 11 (March, 2009), 2098--105.
[7]
Gong, A., and Liu, Y. 2011. Improved KNN Classification Algorithm by Dynamic Obtaining K. International Conference, ECWAC. (April, 2011), 320--324.
[8]
Lin, S. W., and Chen, S. C. 2011. Parameter tuning, feature selection and weight assignment of features for case-based reasoning by artificial immune system. Applied Soft Computing. 11, 8 (December, 2011), 5042--5052.
[9]
Guyon, I., and Elisseeff, Andre. 2002. An introduction to variable and feature selection. J Mach Learn Res. 3, 6 (March, 2002), 1157--1182.
[10]
Saeys, Y., Inza, I., and Larrañaga, P. 2007. A review of feature selection techniques in bioinformatics. Bioinformatics. 23, 19 (August, 2007), 2507--17.
[11]
Chuang, L. Y., Chang, H. W., Tu, C. J., and Yang, C. H. 2008. Improved binary pso for feature selection using gene expression data. Comput Biol Chem. 32, 1 (February, 2008), 29--38.
[12]
Sahu, B., and Mishra, D. 2012. A novel feature selection algorithm using particle swarm optimization for cancer microarray data. Procedia Engineering. 38, 5 (December, 2012), 27--31.
[13]
Kardan, A. A., Kavian, A., and Esmaeili, A. 2013. Simultaneous feature selection and feature weighting with K selection for KNN classification using BBO algorithm. International Conference, IKT. 28, 4, (May, 2013), 349--354.
[14]
Kalayeh, M. M., Idrees, H., and Shah, M. 2014. NMF-KNN: Image Annotation Using Weighted Multi-view Non-negative Matrix Factorization. IEEE Conference on CVPR. (June, 2014), 184--191.
[15]
Wu, J. L., and Li, I. J. 2010. A SOM-based dimensionality reduction method for KNN classifiers. International Conference on SSE. (August, 2010), 173--178.
[16]
Draszawka, K., Szymański, J., and Guerra, F. 2015. Improving css-KNN Classification Performance by Shifts in Training Data. International Conference on IKC. (January, 2016), 51--63.
[17]
Omranpour, H., and Ghidary, S. S. 2016. A heuristic supervised euclidean data difference dimension reduction for knn classifier and its application to visual place classification. Neural Comput & Applic. (October, 2016), 1867--1881.
[18]
Jing, Y., Gou, H., and Zhu, Y. 2013. An Improved Density-Based Method for Reducing Training Data in KNN. International Conference on ICCIS. (January, 2013), 972--975.
[19]
Pawlovsky, A. P., and Nagahashi, M. 2014. A method to select a good setting for the kNN algorithm when using it for breast cancer prognosis. International Conference on BHI. (June, 2014), 189--192.
[20]
Li, N., Kong, H., Ma, Y., Gong, G., and Huai, W. 2016. Human performance modeling for manufacturing based on an improved knn algorithm. Int J Adv Manuf Technol. 84,1 (February, 2016), 473--483.
[21]
Wang, B., Liao, Q., and Zhang, C. 2013. Weight Based KNN Recommender System. International Conference on IHMSC. 2, (August, 2013), 449--452.
[22]
Zhang, L., Zhang, C., Xu, Q., and Liu, C. 2015. Weigted-KNN and its application on UCI. IEEE International Conference on IA. (August, 2015), 1748--1750.
[23]
Shi, K., Li, L., Liu, H., and He, J. 2011. An improved KNN text classification algorithm based on density. IEEE Proceedings of IEEE CCIS. (September, 2011), 113--117.
[24]
Liu, X., Ren, F., and Yuan, C. 2010. Use relative weight to improve the kNN for unbalanced text category. Proceedings of IEEE NLPKE. (August, 2010), 1--5.
[25]
Haixiang, G., Yijing, L., Yanan, L., Xiao, L., and Jinling, L. 2016. Bpso-adaboost-knn ensemble learning algorithm for multi-class unbalanced data classification. Eng Appl Artif Intel, (March, 2016), 176--193.
[26]
Bhattacharya, G., Ghosh, K., and Chowdhury, A. S. 2015. A probabilistic framework for dynamic k estimation in kNN classifiers with certainty factor. Proceedings of IEEE ICAPR. (January, 2015), 1--5.
[27]
Zhang, S., Zong, M., Sun, K., Liu, Y., and Cheng, D. 2014. Efficient kNN Algorithm Based on Graph Sparse Reconstruction. International Conference on ADMA. (December, 2014), 356--369.
[28]
Hamerly, G., and Speegle, G. 2010. Efficient Model Selection for Large-Scale Nearest-Neighbor Data Mining. International Conference on BNCOD 27. (July, 2010), 37--54.
[29]
Weng, F., Jiang, Q., Chen, L., Hong, Z., and Jiang, Q. S. 2007. Clustering ensemble based on the fuzzy KNN algorithm. International Conference on ACIS 18. (July, 2007), 1001--1006.
[30]
Jiang, L., Cai, Z., Wang, D., and Zhang, H. 2014. Bayesian citation-knn with distance weighting. Int. J. Mach. Learn. & Cyber. 5, 2 (April, 2014), 193--199.
[31]
Kozak, K., Kozak, M., Stapor, K., Kozak, K., Kozak, M., and Stapor, K. 2006. Weighted k-nearest-neighbor techniques for high throughput screening data. Int. J. of Bio & Med Sci. (January, 2006), 1--4.
[32]
Loftsgaarden, D. O., and Quesenberry, C. P. 1965. A nonparametric estimate of a multivariate density function. Ann. Math. Statist. 36, 3 (1965), 1049--1051.
[33]
Mitra, P., Murthy, C. A., and Pal, S. K. 2002. Unsupervised feature selection using feature similarity. IEEE Transactions on PAMI. 24, 3 (March, 2002), 301--312.
[34]
Kosinski, M., Matz, S. C., Gosling, S. D., Popov, V., and Stillwell, D. 2015. Facebook as a research tool for the social sciences: opportunities, challenges, ethical considerations, and practical guidelines. American Psychologist. 70, 6 (2015), 543--556.
[35]
Fernández-Tobíasd, I., and Cantador, I. 2014. Personality-Aware Collaborative Filtering: An Empirical Study in Multiple Domains with Facebook Data. International Conference on EC-Web. (September, 2014), 125--137.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICMLC '17: Proceedings of the 9th International Conference on Machine Learning and Computing
February 2017
545 pages
ISBN:9781450348171
DOI:10.1145/3055635
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

In-Cooperation

  • Southwest Jiaotong University

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 24 February 2017

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. classification algorithm
  2. dynamic K
  3. k-Nearest neighbors

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ICMLC 2017

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)24
  • Downloads (Last 6 weeks)0
Reflects downloads up to 16 Oct 2024

Other Metrics

Citations

Cited By

View all

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media