Spoken language acquisition via human-robot interaction

This paper presents a subproject of a challenging project that explores teaching a computer human-intelligence. In the subproject, a multisensory mobile robot is used as the interface for human-computer interaction, and spoken language is taught to the computer through natural human-robot interaction. Different from state-of-the-art speech recognizers, our approach associates speech patterns directly with sensory inputs of the robot. This approach allows our system to learn multilingual speech patterns online. Further investigation of this project will include human-computer interaction that involves more modalities, and applications that use the proposed idea to train home appliances.

[1]  Ying Wu,et al.  Interactive and Incremental Learning via a Mixture of Supervised and Unsupervised Learning Strategies , 2000 .

[2]  A.L. Gorin,et al.  An experiment in spoken language acquisition , 1992, IEEE Trans. Speech Audio Process..

[3]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.

[4]  Stephen E. Levinson,et al.  Language as part of sensorimotor behavior , 1995 .

[5]  R. Gray,et al.  Vector quantization , 1984, IEEE ASSP Magazine.

[6]  Richard O. Duda,et al.  Pattern classification and scene analysis , 1974, A Wiley-Interscience publication.

[7]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[8]  Computing Machinery and Intelligence Mind Vol. 59 , 2022 .

[9]  Alex Pentland,et al.  Learning words from natural audio-visual input , 1998, ICSLP.

[10]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[11]  Teuvo Kohonen,et al.  Self-Organizing Maps , 2010 .

[12]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.