The Proposal for Compensation to the Action of Motion Control based on the Prediction of State-action Pair

For a robot that works in a dynamic environment, the ability to autonomously cope with the changes in the environment, is important. In this paper, an approach to predict the changes of the state and action of the robot is proposed. Further, to extend this approach, the action to be taken in the future will be attempted to apply, to the current action. This method predicts the robot state and action for the distant future using the state that the robot adopts repeatedly. Using this method, the actions that the robot will take in the future can be predicted. In this paper, a method that predicts the state and the action each time the robot decides to perform an action will be proposed. In particular, in this paper, how to define the weight coefficients will be focused on, using the characteristics of the future prediction results. Using this method, the compensatory current action will be obtained. This paper presents the results of our study and discusses methods that allow the robot to decide its desirable behavior quickly, using state prediction and optimal control methods.

[1]  Kentarou Kurashige,et al.  Real-time sequentially decision for optimal action using prediction of the state-action pair , 2014, 2014 International Symposium on Micro-NanoMechatronics and Human Science (MHS).

[2]  Bingrong Hong,et al.  Active Mobile Robot Simultaneous Localization and Mapping , 2006, 2006 IEEE International Conference on Robotics and Biomimetics.

[3]  Tetsuo Ono,et al.  Development of Robovie as a Platform for Everyday-Robot Research , 2002 .

[4]  B. Kuipers,et al.  Robot Navigation with MPEPC in Dynamic and Uncertain Environments : From Theory to Practice , 2012 .

[5]  Fabrizio Abrate,et al.  Multi-robot Map Updating in Dynamic Environments , 2010, DARS.

[6]  P. Protzel,et al.  Using the Unscented Kalman Filter in Mono-SLAM with Inverse Depth Parametrization for Autonomous Airship Control , 2007, 2007 IEEE International Workshop on Safety, Security and Rescue Robotics.

[7]  Mohammed Faisal,et al.  Fuzzy Logic Navigation and Obstacle Avoidance by a Mobile Robot in an Unknown Dynamic Environment , 2013 .

[8]  Nicolas Schweighofer,et al.  Local Online Support Vector Regression for Learning Control , 2007, 2007 International Symposium on Computational Intelligence in Robotics and Automation.

[9]  Mohammad Ali Badamchizadeh,et al.  Extended and Unscented Kalman Filtering Applied to a Flexible-Joint Robot with Jerk Estimation , 2010 .

[10]  Foudil Abdessemed,et al.  SVM-Based Control System for a Robot Manipulator , 2012 .

[11]  Ellips Masehian,et al.  Sensor-Based Motion Planning of Wheeled Mobile Robots in Unknown Dynamic Environments , 2013, Journal of Intelligent & Robotic Systems.

[12]  Hiroshi Ishiguro,et al.  Mobile Robot Navigation by a Distributed Vision System , 1999 .

[13]  H. Jin Kim,et al.  Model predictive flight control using adaptive support vector regression , 2010, Neurocomputing.

[14]  Giancarlo Marafioti,et al.  State Estimation in Nonlinear Model Predictive Control, Unscented Kalman Filter Advantages , 2009 .

[15]  Mohammad A. Jaradat,et al.  Reinforcement based mobile robot navigation in dynamic environment , 2011 .

[16]  Francesco Parrella Online Support Vector Regression , 2007 .

[17]  Blake Hannaford,et al.  Robustness of the Unscented Kalman filter for state and parameter estimation in an elastic transmission , 2009, Robotics: Science and Systems.

[18]  Kentarou Kurashige,et al.  The proposal for deciding effective action using prediction of internal robot state based on internal state and action , 2013, MHS2013.

[19]  Thomas J. Walsh,et al.  Planning and Learning in Environments with Delayed Feedback , 2007, ECML.

[20]  Minoru Asada,et al.  Purposive Behavior Acquisition for a Robot by Vision-Based Reinforcement Learning , 1995 .

[21]  M Abdel Kareem,et al.  REINFORCEMENT BASED MOBILE ROBOT NAVIGATION IN DYNAMIC ENVIRONMENT , 2011 .

[22]  Wolfram Burgard,et al.  Probabilistic Robotics (Intelligent Robotics and Autonomous Agents) , 2005 .

[23]  Shunichi Asaka,et al.  Behavior Control of an Autonomous Mobile Robot in Dynamically Changing Environment , 1994 .

[24]  Chih-Jen Lin,et al.  Training and Testing Low-degree Polynomial Data Mappings via Linear SVM , 2010, J. Mach. Learn. Res..

[25]  Minoru Asada,et al.  Incremental State Space Segmentation for Behavior Learning by Real Robot , 1999 .

[26]  W. Burgard,et al.  Markov Localization for Mobile Robots in Dynamic Environments , 1999, J. Artif. Intell. Res..

[27]  Robert Babuska,et al.  Control delay in Reinforcement Learning for real-time dynamic systems: A memoryless approach , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[28]  Kok Kiong Tan,et al.  Computation delay compensation for real time implementation of robust model predictive control , 2012, IEEE 10th International Conference on Industrial Informatics.

[29]  Gaurav S. Sukhatme,et al.  Mobile Robot Simultaneous Localization and Mapping in Dynamic Environments , 2005, Auton. Robots.

[30]  Isao Ohmura,et al.  2A2-C10 A Study on Model-Based Development of Embedded System using Scilab/Scicos : Development of Auto-Code Generator , 2010 .

[31]  Mohammad Teshnehlab,et al.  Adaptive Neuro-Fuzzy Extended Kaiman Filtering for robot localization , 2010, Proceedings of 14th International Power Electronics and Motion Control Conference EPE-PEMC 2010.

[32]  J. G. Iossaqui SLIP ESTIMATION USING THE UNSCENTED KALMAN FILTER FOR THE TRACKING CONTROL OF MOBILE ROBOTS , 2011 .

[33]  Wen-Hua Chen,et al.  Model predictive control for autonomous helicopters with computational delay , 2010 .