首页 | 本学科首页   官方微博 | 高级检索  
     


Self-refreshing memory in artificial neural networks: learning temporal sequences without catastrophic forgetting
Authors:Bernard Ans  Stéphane Rousset  Robert M. French  Serban Musca
Affiliation:1. Psychology and NeuroCognition , Pierre Mendes-France University , Grenoble 2—CNRS UMR 5105, BP 47, 38040, Grenoble cedex 09, France Phone: +33-476-825-674 Fax: +33-476-825-674 E-mail: Bernard.Ans@upmf-grenoble.fr;2. Quantitative Psychology and Cognitive Science , University of Liège , (Bat B32), Sart Tilman, 4000, Liège, Belgium
Abstract:While humans forget gradually, highly distributed connectionist networks forget catastrophically: newly learned information often completely erases previously learned information. This is not just implausible cognitively, but disastrous practically. However, it is not easy in connectionist cognitive modelling to keep away from highly distributed neural networks, if only because of their ability to generalize. A realistic and effective system that solves the problem of catastrophic interference in sequential learning of ‘static’ (i.e. non-temporally ordered) patterns has been proposed recently (Robins 1995 Robins, AV. 1995. Catastrophic forgetting, rehearsal and pseudorehearsal. Connection Science, 7: 123146. [Taylor & Francis Online] [Google Scholar], Connection Science, 7: 123–146, 1996, Connection Science, 8: 259–275, Ans and Rousset 1997 Ans, B and Rousset, S. 1997. Avoiding catastrophic forgetting by coupling two reverberating neural networks. CR Académie des Sciences Paris, Life Sciences, 320: 989997.  [Google Scholar], CR Académie des Sciences Paris, Life Sciences, 320: 989–997, French 1997 French, RM. 1997. Pseudo-recurrent connectionist networks: an approach to the ‘sensitivity–stability’ dilemma. Connection Science, 9: 353379. [Taylor & Francis Online] [Google Scholar], Connection Science, 9: 353–379, 1999, Trends in Cognitive Sciences, 3: 128–135, Ans and Rousset 2000 Ans, B and Rousset, S. 2000. Neural networks with a self-refreshing memory: knowledge transfer in sequential learning tasks without catastrophic forgetting. Connection Science, 12: 119. [Taylor & Francis Online], [Web of Science ®] [Google Scholar], Connection Science, 12: 1–19). The basic principle is to learn new external patterns interleaved with internally generated ‘pseudopatterns’ (generated from random activation) that reflect the previously learned information. However, to be credible, this self-refreshing mechanism for static learning has to encompass our human ability to learn serially many temporal sequences of patterns without catastrophic forgetting. Temporal sequence learning is arguably more important than static pattern learning in the real world. In this paper, we develop a dual-network architecture in which self-generated pseudopatterns reflect (non-temporally) all the sequences of temporally ordered items previously learned. Using these pseudopatterns, several self-refreshing mechanisms that eliminate catastrophic forgetting in sequence learning are described and their efficiency is demonstrated through simulations. Finally, an experiment is presented that evidences a close similarity between human and simulated behaviour.
Keywords:artificial neural networks  temporal sequence learning  catastrophic forgetting  self-refreshing memory
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号