Generation of Comprehensible Representations by Supposed Maximum Information

In this paper, we propose a new information-theoretic method to simplify and unify learning methods in one framework. The new method is called "supposed maximum information," which is used to produce humanly comprehensible internal representations supposing that information is already maximized before learning. The new learning method is composed of three stages. First, without information on input variables, a competitive network is trained. Second, with supposed maximum information on input variables, the importance of the variables is estimated by measuring mutual information between competitive units and input patterns. Finally, with the estimated importance of input variables, the competitive network is retrained to take into account the importance of input variables. The method is applied not to pure competitive learning but to self-organizing maps, because it is easy to demonstrate how well the new method can produce more explicit internal representation intuitively. We applied the method to the well-known SPECT heart data of the machine learning database. We succeeded in producing more comprehensible class boundaries on the U-matrices than did the conventional SOM. In addition, quantization and topographic errors produced by our method were not larger than those by the conventional SOM.

[1]  Suzanna Becker,et al.  Mutual information maximization: models of cortical self-organization. , 1996, Network.

[2]  Kari Torkkola,et al.  Feature Extraction by Non-Parametric Mutual Information Maximization , 2003, J. Mach. Learn. Res..

[3]  H. B. Barlow,et al.  Unsupervised Learning , 1989, Neural Computation.

[4]  Ralph Linsker,et al.  Self-organization in a perceptual network , 1988, Computer.

[5]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[6]  Teuvo Kohonen,et al.  Self-Organizing Maps , 2010 .

[7]  Ryotaro Kamimura,et al.  Information-Theoretic Competitive Learning with Inverse Euclidean Distance Output Units , 2003, Neural Processing Letters.

[8]  Kimmo Kiviluoto,et al.  Topology preservation in self-organizing maps , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).

[9]  David Zipser,et al.  Feature Discovery by Competive Learning , 1986, Cogn. Sci..

[10]  Rui Xu,et al.  Survey of clustering algorithms , 2005, IEEE Transactions on Neural Networks.

[11]  Juha Vesanto,et al.  SOM-based data visualization methods , 1999, Intell. Data Anal..

[12]  M. V. Velzen,et al.  Self-organizing maps , 2007 .

[13]  John W. Fisher,et al.  Learning from Examples with Information Theoretic Criteria , 2000, J. VLSI Signal Process..