Self-Learning for a Humanoid Robotic Ping-Pong Player

Imitating the learning process of a human playing ping-pong is extremely complex. This work proposes a suitable learning strategy. First, an inverse kinematics solution is presented to obtain the smooth joint angles of a redundant anthropomorphic robot arm in order to imitate the paddle motion of a human ping-pong player. As humans instinctively determine which posture is suitable for striking a ball, this work proposes two novel processes: (i) estimating ball states and predicting trajectory using a fuzzy adaptive resonance theory network, and (ii) self-learning the behavior for each strike using a self-organizing map-based reinforcement learning network that imitates human learning behavior. Experimental results demonstrate that the proposed algorithms work effectively when applied to an actual humanoid robot playing ping-pong.

[1]  Fumio Miyazaki,et al.  A learning approach to robotic table tennis , 2005, IEEE Transactions on Robotics.

[2]  Fuminori Saito,et al.  Learning architecture for real robotic systems-extension of connectionist Q-learning for continuous robot control domain , 1994, Proceedings of the 1994 IEEE International Conference on Robotics and Automation.

[3]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[4]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[5]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[6]  Fan-Tien Cheng,et al.  Window-shaped obstacle avoidance for a redundant manipulator , 1998, IEEE Trans. Syst. Man Cybern. Part B.

[7]  Danchi Jiang,et al.  A Lagrangian network for kinematic control of redundant robot manipulators , 1999, IEEE Trans. Neural Networks.

[8]  C. Touzet,et al.  Self-organizing map for reinforcement learning: obstacle-avoidance with Khepera , 1994, Proceedings of PerAc '94. From Perception to Action.

[9]  Christopher G. Atkeson,et al.  A Framework for Learning from Observation Using Primitives , 2002, RoboCup.

[10]  C. H. Lai,et al.  Design and hand-eye coordination of a humanoid robot , 2008 .

[11]  Long Ji Lin,et al.  Self-improving reactive agents based on reinforcement learning, planning and teaching , 1992, Machine Learning.

[12]  Juan A. Méndez,et al.  Ping-pong player prototype , 2003, IEEE Robotics Autom. Mag..

[13]  C. S. G. Lee,et al.  Robotics: Control, Sensing, Vision, and Intelligence , 1987 .

[14]  Stephen Grossberg,et al.  Fuzzy ARTMAP: A neural network architecture for incremental supervised learning of analog multidimensional maps , 1992, IEEE Trans. Neural Networks.

[15]  Zhengyou Zhang,et al.  A Flexible New Technique for Camera Calibration , 2000, IEEE Trans. Pattern Anal. Mach. Intell..