On the Probabilistic Interpretation of Neural Network Behavior

Recent probabilistic interpretations of neural network models have suggested the formulaton of network operations in information-theoretic terms. In these interpretations, the neural network developes an assumed probability density function which represents its assumptions on the environment. Using a set of hypotheses, this probability density functon is shown to maintain an exponential relationship wth an energy-like functon that the network tends to minimize. The purpose of this note is to obtain this probability density function through Shannon's dervation of the entropy measure and Jaynes' maximum entropy principle. The main conclusion is that the neural network assumes the worst case (i.e. most uncertain or maximum-entropy) probability density function for the unknown environment.

[1]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[2]  Allon Guez,et al.  On the stability, storage capacity, and design of nonlinear continuous neural networks , 1988, IEEE Trans. Syst. Man Cybern..

[3]  E. Jaynes Information Theory and Statistical Mechanics , 1957 .

[5]  J J Hopfield,et al.  Neurons with graded response have collective computational properties like those of two-state neurons. , 1984, Proceedings of the National Academy of Sciences of the United States of America.

[6]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.