A New $k$-Nearest Neighbors Algorithm for Learning from Multiple Experts' Uncertain Decisions

$k$-nearest neighbors algorithm ($k$-NN) is a simple yet powerful method that can predict an unlabeled object using the $k$ closest samples whose labels are already known. However, in some cases such as medical treatment decision making, multiple experts can provide their individual decisions and there is no single ground truth for the labels of samples. As a result, the accuracy of the predictions can be affected by the uncertainties of historical decisions. In order to learn from uncertain decisions from multiple experts, we propose a new $k$-NN algorithm. This algorithm firstly measures the uncertainties of labels. Then, the prediction is made based on the uncertain labels of the $k$ closest samples based on the Dempster-Shafer theory. We also introduce two methods to compute the optimal $k$ value. Experiments on a real-world medical treatment dataset prove that our algorithm has better performance compared with other algorithms.

[1]  Md. Rafiqul Islam,et al.  Medical image classfication using an efficient data mining technique , 2004, 2004 International Conference on Machine Learning and Applications, 2004. Proceedings..

[2]  Nicolás García-Pedrajas,et al.  A Proposal for Local $k$ Values for $k$ -Nearest Neighbor Rule , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[3]  Byoung-Tak Zhang,et al.  Generative Local Metric Learning for Nearest Neighbor Classification , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Jinkun Chen,et al.  A new nearest neighbor classifier via fusing neighborhood information , 2014, Neurocomputing.

[5]  Jian Cao,et al.  A Multi-disciplinary Medical Treatment Decision Support System with intelligent treatment recommendation , 2016, 2016 2nd IEEE International Conference on Computer and Communications (ICCC).

[6]  Yoav Freund,et al.  Improved kNN Rule for Small Training Sets , 2014, 2014 13th International Conference on Machine Learning and Applications.

[7]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[8]  Xuelong Li,et al.  Learning k for kNN Classification , 2017, ACM Trans. Intell. Syst. Technol..

[9]  Philip S. Yu,et al.  Top 10 algorithms in data mining , 2007, Knowledge and Information Systems.

[10]  R. Gray Entropy and Information Theory , 1990, Springer New York.

[11]  Beng Chin Ooi,et al.  Indexing the Distance: An Efficient Method to KNN Processing , 2001, VLDB.

[12]  Bhavani M. Thuraisingham,et al.  An Effective Evidence Theory Based K-Nearest Neighbor (KNN) Classification , 2008, 2008 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology.

[13]  Thierry Denoeux,et al.  A k-nearest neighbor classification rule based on Dempster-Shafer theory , 1995, IEEE Trans. Syst. Man Cybern..

[14]  Ye Shi SVM-KNN Classifier——A New Method of Improving the Accuracy of SVM Classifier , 2002 .

[15]  A. Matarasso Abdominolipoplasty: A system of classfication and treatment for combined abdominoplasty and suction-assisted lipectomy , 2005, Aesthetic Plastic Surgery.

[16]  Ming-Syan Chen,et al.  DIKNN: An Itinerary-based KNN Query Processing Algorithm for Mobile Sensor Networks , 2007, 2007 IEEE 23rd International Conference on Data Engineering.

[17]  Xuelong Li,et al.  Block-Row Sparse Multiview Multilabel Learning for Image Classification , 2016, IEEE Transactions on Cybernetics.