Parallel and local learning for fast probabilistic neural networks in scalable data mining

Probabilistic Neural Networks (PNN) represent knowledge in the form of simple well understood Bayesian models, which are most suitable for data mining applications that also need confidence levels. However if the number N of pattern neurons grows large, the evaluation of a single unknown sample has still O(N) complexity, which is not scale well. Parallelism can efficiently work towards speeding up the large neural networks. Whereas the PNN run time can be reduced by parallelism, its computational cost can be decreased by keeping only the globally most important pattern neurons. While such parsimonious global models have been studied for a long time, there is another way that is related to local learning algorithms. Maintain all N pattern neurons but to classify any unknown x use only the local ones, the k nearest neighbor to x neurons. This local learning PNN is investigated here. By using confidence ratio outputs we optimize the number of k nearest neighbor neurons for the best PNN performance. Combined with the parallel approach using p processors the method can reduce the cost of classifying an unknown x from O(N) to O(k/p).

[1]  Roy L. Streit,et al.  Maximum likelihood training of probabilistic neural networks , 1994, IEEE Trans. Neural Networks.

[2]  Michael Georgiopoulos,et al.  Methods for Parallelizing the Probabilistic Neural Network on a Beowulf Cluster Computer , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[3]  Michael N. Vrahatis,et al.  Novel Approaches to Probabilistic Neural Networks Through Bagging and Evolutionary Estimating of Prior Probabilities , 2008, Neural Processing Letters.

[4]  Pietro Burrascano,et al.  Learning vector quantization for the probabilistic neural network , 1991, IEEE Trans. Neural Networks.

[5]  Léon Bottou,et al.  Local Learning Algorithms , 1992, Neural Computation.

[6]  Wee Ser,et al.  Probabilistic neural-network structure determination for pattern classification , 2000, IEEE Trans. Neural Networks Learn. Syst..

[7]  Octavia I. Camps,et al.  Weighted Parzen Windows for Pattern Classification , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  D. F. Specht,et al.  Experience with adaptive probabilistic neural networks and adaptive general regression neural networks , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[9]  E. Parzen On Estimation of a Probability Density Function and Mode , 1962 .

[10]  Bernhard Schölkopf,et al.  A Local Learning Approach for Clustering , 2006, NIPS.

[11]  Enrico Blanzieri,et al.  Fast and Scalable Local Kernel Machines , 2010, J. Mach. Learn. Res..

[12]  Sundaram Suresh,et al.  Parallel implementation of back-propagation algorithm in networks of workstations , 2005, IEEE Transactions on Parallel and Distributed Systems.

[13]  Léon Bottou,et al.  Local Algorithms for Pattern Recognition and Dependencies Estimation , 1993, Neural Computation.

[14]  Donald F. Specht,et al.  Probabilistic neural networks , 1990, Neural Networks.

[15]  Chu Kiong Loo,et al.  A Global k-means Approach for Autonomous Cluster Initialization of Probabilistic Neural Network , 2008, Informatica.

[16]  Nikola B. Serbedzija Simulating Artificial Neural Networks on Parallel Architectures , 1996, Computer.