On the Theory and Applications of Sequence Based Estimation of Independent Binomial Random Variables

We re-visit the age-old problem of estimating the parameters of a distribution from its observations. Traditionally, scientists and statisticians have attempted to obtain strong estimates by “extracting” the information contained in the observations taken as a set. However, generally speaking, the information contained in the sequence in which the observations have appeared, has been ignored – i.e., except to consider dependence information as in the case of Markov models and n-gram statistics. In this paper, we present results which, to the best of our knowledge, are the first reported results, which consider how estimation can be enhanced by utilizing both the information in the observations and in their sequence of appearance. The strategy, known as Sequence Based Estimation (SBE) works as follows. We first quickly allude to the results pertaining to computing the Maximum Likelihood Estimates (MLE) of the data when the samples are taken individually. We then derive the corresponding MLE results when the samples are taken two-at-a-time, and then extend these for the cases when they are processed three-at-a-time, four-at-a-time etc. In each case, we also experimentally demonstrate the convergence of the corresponding estimates. We then suggest various avenues for future research, including those by which these estimates can be fused to yield a superior overall cumulative estimate of the parameter of the distribution. We believe that our new estimates have great potential for practitioners, especially when the cardinality of the observation set is small.

[1]  Ralf Herbrich,et al.  Learning Kernel Classifiers: Theory and Algorithms , 2001 .

[2]  David G. Stork,et al.  Pattern Classification (2nd ed.) , 1999 .

[3]  Dilip D. Kandlur,et al.  Adaptive packet marking for maintaining end-to-end throughput in a differentiated-services internet , 1999, TNET.

[4]  P. Bickel,et al.  Mathematical Statistics: Basic Ideas and Selected Topics , 1977 .

[5]  David G. Stork,et al.  Pattern Classification , 1973 .

[6]  Andrew R. Webb,et al.  Statistical Pattern Recognition , 1999 .

[7]  Andrew S. Tanenbaum,et al.  Computer Networks , 1981 .

[8]  Glen G. Langdon,et al.  An Introduction to Arithmetic Coding , 1984, IBM J. Res. Dev..

[9]  Ludmila I. Kuncheva,et al.  A Theoretical Study on Six Classifier Fusion Strategies , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  G. Casella,et al.  Statistical Inference , 2003, Encyclopedia of Social Network Analysis and Mining.

[11]  B. John Oommen,et al.  On the estimation of independent binomial random variables using occurrence and sequential information , 2007, Pattern Recognit..

[12]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  P. P. White,et al.  RSVP and integrated services in the Internet: a tutorial , 1997, IEEE Commun. Mag..

[14]  David G. Stork,et al.  Pattern classification, 2nd Edition , 2000 .

[15]  Glen G. Langdon,et al.  Arithmetic Coding , 1979 .

[16]  Paul H. Garthwaite,et al.  Statistical Inference , 2002 .

[17]  Ian H. Witten,et al.  Arithmetic coding for data compression , 1987, CACM.

[18]  Xipeng Xiao,et al.  Internet QoS: a big picture , 1999, IEEE Netw..

[19]  A Gordon,et al.  Classification, 2nd Edition , 1999 .

[20]  Horst Bunke Structural and Syntactic Pattern Recognition , 1993, Handbook of Pattern Recognition and Computer Vision.