Generalization by symbolic abstraction in cascaded recurrent networks
暂无分享,去创建一个
[1] Mikael Bodén,et al. Learning the Dynamics of Embedded Clauses , 2003, Applied Intelligence.
[2] Dirk V. Arnold,et al. Syntactic systematicity arising from semantic predictions in a Hebbian-competitive network , 2001, Connect. Sci..
[3] Gary F Marcus,et al. Language acquisition in the absence of explicit negative evidence: can simple recurrent networks obviate the need for domain-specific learning devices? , 1999, Cognition.
[4] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[5] Steven Phillips,et al. Constituent similarity and systematicity: The limits of first-order connectionism , 2000, Connect. Sci..
[6] Douglas L. T. Rohde,et al. Simple recurrent networks can distinguish non-occurring from ungrammatical sentences given appropriate task structure: reply to Marcus , 1999, Cognition.
[7] J. Fodor,et al. Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.
[8] Jürgen Schmidhuber,et al. Discovering Predictable Classifications , 1993, Neural Computation.
[9] Stefano Nolfi,et al. Extracting Regularities in Space and Time Through a Cascade of Prediction Networks: The Case of a Mobile Robot Navigating in a Structured Environment , 1999, Connect. Sci..
[10] Janet Wiles,et al. Context-free and context-sensitive dynamics in recurrent neural networks , 2000, Connect. Sci..
[11] J. Pollack. The Induction of Dynamical Recognizers , 1996, Machine Learning.
[12] Robert F. Hadley. Systematicity in Connectionist Language Learning , 1994 .
[13] Nick Chater,et al. Toward a connectionist model of recursion in human linguistic performance , 1999 .
[14] Zenon W. Pylyshyn,et al. Connectionism and cognitive architecture , 1993 .
[15] Paul Rodríguez,et al. A Recurrent Neural Network that Learns to Count , 1999, Connect. Sci..
[16] Douglas L. T. Rohde,et al. Language acquisition in the absence of explicit negative evidence: how important is starting small? , 1999, Cognition.
[17] Peter M. Vishton,et al. Rule learning by seven-month-old infants. , 1999, Science.
[18] Jürgen Schmidhuber,et al. LSTM recurrent networks learn simple context-free and context-sensitive languages , 2001, IEEE Trans. Neural Networks.
[19] J. Elman. Learning and development in neural networks: the importance of starting small , 1993, Cognition.
[20] Ron Kohavi,et al. Supervised and Unsupervised Discretization of Continuous Features , 1995, ICML.