Position Paper: Representation Search through Generate and Test

Learning representations from data is one of the fundamental problems of artificial intelligence and machine learning. Many different approaches exist for learning representations, but what constitutes a good representation is not yet well understood. In this work, we view the problem of representation learning as one of learning features (e.g., hidden units of neural networks) such that performance of the underlying base system continually improves. We study an important case where learning is done fully online (i.e., on an example-by-example basis) from an unending stream of data, and the computational cost of the learning element should not grow with time or cannot be much more than that of the performance element. Few methods can be used effectively in this case. We show that a search approach to representation learning can naturally fit with this setting. In this approach good representations are searched by generating different features and then testing them for utility. We develop new representation-search methods and show that the generate-and-test approach can be utilized in a simple and effective way for continually improving representations. Our methods are fully online and add only a small fraction to the overall computation. We believe online representation search constitutes an important step toward effective and inexpensive solutions to representation learning problems.

[1]  Charles W. Anderson,et al.  Q-Learning with Hidden-Unit Restarting , 1992, NIPS.

[2]  Peter Vamplew,et al.  Global Versus Local Constructive Function Approximation for On-Line Reinforcement Learning , 2005, Australian Conference on Artificial Intelligence.

[3]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[4]  Pascal Vincent,et al.  Unsupervised Feature Learning and Deep Learning: A Review and New Perspectives , 2012, ArXiv.

[5]  Ron Kohavi,et al.  Irrelevant Features and the Subset Selection Problem , 1994, ICML.

[6]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[7]  Shimon Whiteson,et al.  Evolutionary Function Approximation for Reinforcement Learning , 2006, J. Mach. Learn. Res..

[8]  Russell Reed,et al.  Pruning algorithms-a survey , 1993, IEEE Trans. Neural Networks.

[9]  Pierre Comon Independent component analysis - a new concept? signal processing , 1994 .

[10]  Richard S. Sutton,et al.  Online Learning with Random Representations , 1993, ICML.

[11]  Pierre Comon,et al.  Independent component analysis, A new concept? , 1994, Signal Process..

[12]  Jihoon Yang,et al.  Constructive neural-network learning algorithms for pattern classification , 2000, IEEE Trans. Neural Networks Learn. Syst..

[13]  David Ha Evolutionary Function Approximation for Reinforcement Learning , 2015 .

[14]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[15]  A. Klopf,et al.  An Evolutionary Pattern Recognition Network , 1969 .

[16]  Christian Lebiere,et al.  The Cascade-Correlation Learning Architecture , 1989, NIPS.