Life-long Learning Through Task Rehearsal and Selective Knowledge Transfer

The majority of machine learning research has focused on the single task learning (STL) approach where an hypothesis for a single task is induced from a set of supervised training examples. In contrast, one of the key aspects of human learning is that individuals face a sequence of learning problems over a lifetime. Humans take advantage of this by transferring knowledge from previously learned tasks to facilitate the learning of a new task. Life-long learning, a relatively new area of machine learning research, is concerned with the persistent and cumulative nature of learning (Thrun, 1997). Life-long learning considers situations in which a learner faces a series of different tasks and develops methods of retaining and using prior knowledge to improve the effectiveness (more accurate hypotheses) and efficiency (shorter training times) of learning. Related names for life-long learning in the literature are learning to learn and meta-learning. A challenge often faced by a life-long learning agent is a deficiency of training examples from which to develop accurate hypotheses. Machine learning theory tells us that this problem can be overcome with an appropriate inductive bias (Mitchell, 1997), one source being prior task knowledge (Baxter, 1995). Lacking a theory of knowledge transfer (Caruana, 1997, Thrun, 1997) that distinguishes knowledge from related and unrelated tasks, we have developed one and applied it to life-long learning problems, such as learning a more accurate medical diagnostic model from a small sample of patient data (Silver, 2000). The approach requires (1) a method of selectively transferring previously learned knowledge to a new task based on a measure of task relatedness and (2) a method of retaining learned task knowledge and its recall when learning a new task. In (Silver & Mercer, 1996) we introduced ηMTL, a modified version of the multiple task learning (MTL) method of functional transfer to provide a solution to the first problem of selective transfer. Using a measure of previously learned secondary task to primary task relatedness, an ηMTL network can favourably bias the induction of a hypothesis for a primary task. Section 3 reviews the necessary aspects of ηMTL. This paper focuses on the Task Rehearsal Method (TRM) to solve the second problem of retention and recall of learned task knowledge. TRM uses either the standard MTL or the ηMTL learning algorithms as the method of knowledge transfer and inductive bias. Task rehearsal is so named because previously learned tasks are relearned or rehearsed in parallel O pe n A cc es s D at ab as e w w w .in te ch w eb .o rg

[1]  Jude Shavlik,et al.  Refinement ofApproximate Domain Theories by Knowledge-Based Neural Networks , 1990, AAAI.

[2]  Anthony V. Robins,et al.  Consolidation in Neural Networks and in the Sleeping Brain , 1996, Connect. Sci..

[3]  Ricardo Vilalta,et al.  A Perspective View and Survey of Meta-Learning , 2002, Artificial Intelligence Review.

[4]  Rich Caruana,et al.  Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.

[5]  T. Shultz,et al.  Knowledge-based Cascade-correlation: Varying the Size and Shape of Relevant Prior Knowledge , 2003 .

[6]  Sebastian Thrun,et al.  Is Learning The n-th Thing Any Easier Than Learning The First? , 1995, NIPS.

[7]  Satinder Singh Transfer of Learning by Composing Solutions of Elemental Sequential Tasks , 1992, Mach. Learn..

[8]  Meng Joo Er,et al.  Theory and Novel Applications of Machine Learning , 2009 .

[9]  Daniel L. Silver,et al.  Sequential Consolidation of Learned Task Knowledge , 2004, Canadian Conference on AI.

[10]  Sebastian Thrun,et al.  Learning to Learn: Introduction and Overview , 1998, Learning to Learn.

[11]  James L. Carroll,et al.  Reinforcement Learning Task Clustering , 2003, ICMLA.

[12]  Mark B. Ring CHILD: A First Step Towards Continual Learning , 1997, Machine Learning.

[13]  Mark B. Ring Learning Sequential Tasks by Incrementally Adding Higher Orders , 1992, NIPS.

[14]  Ryszard S. Michalski,et al.  Learning = Inferencing + Memorizing , 1993 .

[15]  Anthony V. Robins,et al.  Catastrophic Forgetting, Rehearsal and Pseudorehearsal , 1995, Connect. Sci..

[16]  Richard J. Mammone,et al.  Artificial neural networks for speech and vision , 1994 .

[17]  D. Silver,et al.  Continued Practice and Consolidation of a Learning Task , 2005 .

[18]  P. Ut Goff,et al.  Machine learning of inductive bias , 1986 .

[19]  Stephen José Hanson,et al.  A stochastic version of the delta rule , 1990 .

[20]  Robert E. Mercer,et al.  Selective transfer of neural network task knowledge , 2000 .

[21]  E. Kehoe A layered network model of associative learning: learning to learn and configuration. , 1988, Psychological review.

[22]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[23]  Lorien Y. Pratt,et al.  Discriminability-Based Transfer between Neural Networks , 1992, NIPS.

[24]  R. Detrano,et al.  International application of a new probability algorithm for the diagnosis of coronary artery disease. , 1989, The American journal of cardiology.

[25]  Chuan Wang,et al.  Training neural networks with additive noise in the desired signal , 1999, IEEE Trans. Neural Networks.

[26]  Jonathan Baxter,et al.  Learning internal representations , 1995, COLT '95.

[27]  Stephen Grossberg,et al.  Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..

[28]  Denis J. Dean,et al.  Comparative accuracies of artificial neural networks and discriminant analysis in predicting forest cover types from cartographic variables , 1999 .

[29]  Daniel L. Silver,et al.  The Parallel Transfer of Task Knowledge Using Dynamic Learning Rates Based on a Measure of Relatedness , 1996, Connect. Sci..

[30]  Sebastian Thrun,et al.  Explanation-Based Neural Network Learning for Robot Control , 1992, NIPS.

[31]  Satinder Singh Transfer of learning by composing solutions of elemental sequential tasks , 2004, Machine Learning.

[32]  Robert A. Jacobs,et al.  Increased rates of convergence through learning rate adaptation , 1987, Neural Networks.

[33]  Sebastian Thrun,et al.  Clustering Learning Tasks and the Selective Cross-Task Transfer of Knowledge , 1998, Learning to Learn.

[34]  Michael McCloskey,et al.  Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .

[35]  D. Silver,et al.  Selective Functional Transfer : Inductive Bias from Related Tasks , 2001 .

[36]  Hilbert J. Kappen,et al.  On-line learning processes in artificial neural networks , 1993 .

[37]  Sebastian Thrun,et al.  Lifelong Learning Algorithms , 1998, Learning to Learn.

[38]  Tom M. Mitchell,et al.  The Need for Biases in Learning Generalizations , 2007 .

[39]  Daniel L. Silver,et al.  The Consolidation of Neural Network Task Knowledge , 2003, ICMLA.

[40]  Jude Shavlik,et al.  An Approach to Combining Explanation-based and Neural Learning Algorithms , 1989 .

[41]  Christian Lebiere,et al.  The Cascade-Correlation Learning Architecture , 1989, NIPS.

[42]  A. K. Rigler,et al.  Accelerating the convergence of the back-propagation method , 1988, Biological Cybernetics.

[43]  S. C. Suddarth,et al.  Rule-Injection Hints as a Means of Improving Network Performance and Learning Time , 1990, EURASIP Workshop.

[44]  Mark B. Ring Child: A First Step Towards Continual Learning , 1998, Learning to Learn.

[45]  James L. McClelland,et al.  Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory. , 1995, Psychological review.

[46]  Lorien Y. Pratt,et al.  A Survey of Transfer Between Connectionist Networks , 1996, Connect. Sci..

[47]  Sebastian Thrun,et al.  Lifelong Learning: A Case Study. , 1995 .