Over the years, one of the methods of choice to estimate probability density functions for a given random variable (defined on binary input space) has been the expansion of the estimation function in Rademacher-Walsh Polynomial basis functions. For a set of $L$ features (often considered as an ``$L$-dimensional binary vector''), the Rademacher-Walsh Polynomial approach requires $2^{L}$ basis functions. This can quickly become computationally complicated and notationally clumsy to handle whenever the value of $L$ is large. In current pattern recognition applications it is often the case that the value of $L$ can be 100 or more. In this paper we show that the expansion of the probability density function estimation in Rademacher-Walsh Polynomial basis functions is equivalent to the expansion of the estimation function in a set of Dirac kernel functions. The latter approach is not only able to eloquently allay the computational bottle--neck and notational awkwardness mentioned above, but may also be naturally neater and more ``elegant'' than the Rademacher-Walsh Polynomial basis function approach even when this latter approach is computationally feasible.
[1]
David J. Hand,et al.
Discrimination and Classification
,
1982
.
[2]
William S. Meisel,et al.
Computer-oriented approaches to pattern recognition
,
1972
.
[3]
K. F. Riley,et al.
Mathematical Methods for Physics and Engineering
,
1998
.
[4]
Julius T. Tou,et al.
Pattern Recognition Principles
,
1974
.
[5]
Nello Cristianini,et al.
Kernel Methods for Pattern Analysis
,
2003,
ICTAI.
[6]
Nello Cristianini,et al.
Kernel Methods for Pattern Analysis
,
2004
.
[7]
Jean-Philippe Vert,et al.
Protein-ligand interaction prediction: an improved chemogenomics approach
,
2008,
Bioinform..
[8]
E. Parzen.
On Estimation of a Probability Density Function and Mode
,
1962
.
[9]
N. Aronszajn.
Theory of Reproducing Kernels.
,
1950
.