Home Environment Interaction via Service Robots and theLeap Motion Controller

Ageing society and individuals with disabilities faces numerous challenges in performing simple tasks in Activities of Daily Living (ADLs), [1]. ADLs represent the everyday tasks people usually need to be able to independently accomplish. Nowadays caring of elderly people becomes more and more important. Performance of ADLs in the long term view can be considered as a serious concern, especially when dealing with individuals who require extreme caregiver assistance. The objective of this paper is to introduce an Ambient Intelligence [2], real-scale home environment implementation, embedded with sensors and actuators, which enhances the independence and autonomy of the individuals upon performing ADLs. A real 1:1 scale experimental flat has been design and developed (Figure 1), in the authors experimental laboratory. A study was made to be able to determine the actual needs, required services, and functionality of the proposed augmented environment. In order to allow a high quality service delivery, a mobile autonomous rover [3] was introduced to act as the main humanmachine interface between the user and the distributed robotic systems and actuators. The mobile rover was wirelessly interfaced with the distributed intelligence to allow efficient interaction with the user, either passively, by vocal commands issued by the user, or adaptively, by autonomously navigating into the home environment. Moreover the Leap Motion hand gesture driven controller [4] is used as an intuitive user interface, which allows submillimetre accuracy capabilities, interfaced to a Jaco &-Degrees of freedom robotic arm [5]. Many robotic systems have been designed and produced for assisting ADLs, in order to compensate this loss of mobility. However most of these solution are operated via a keyboard or joystick, which according to the complexity of the robotic system, require a series of configurations and mode selection routines in order to allow a specific trajectory path to be implemented. The authors developed new intuitive interfaces to operate such devices, in order to reduce the involved operation complexity. The user is able with simple gestures to operate and control complex robotic manipulators and mobile robots, without requiring the use of a joystick or a keyboard. The resulting accuracy and evaluated performance of the implemented interfaces, allow error free, continuous, and adaptive interaction between user queries and actuated environment responses.

[1]  Frank Weichert,et al.  Analysis of the Accuracy and Robustness of the Leap Motion Controller , 2013, Sensors.

[2]  Kent Larson,et al.  Using a Live-In Laboratory for Ubiquitous Computing Research , 2006, Pervasive.

[3]  Andreas Stolcke,et al.  DynaSpeak: SRI's scalable speech recognizer for embedded and mobile systems , 2002 .

[4]  José A. Castellanos,et al.  Simultaneous Localization and Map Building , 1999 .

[5]  Alexander I. Rudnicky,et al.  Pocketsphinx: A Free, Real-Time Continuous Speech Recognition System for Hand-Held Devices , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[6]  H.J.A. Stuyt,et al.  Cost-savings and economic benefits due to the assistive robotic manipulator (ARM) , 2005, 9th International Conference on Rehabilitation Robotics, 2005. ICORR 2005..

[7]  Sebastian Stüker,et al.  Rapid porting of ASR-systems to mobile devices , 2005, INTERSPEECH.

[8]  Sander Oude Elberink,et al.  Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications , 2012, Sensors.

[9]  T. Sato,et al.  Environment-type robot system "RoboticRoom" featured by behavior media, behavior contents, and behavior adaptation , 2004, IEEE/ASME Transactions on Mechatronics.

[10]  Philip Requejo,et al.  Mobile Arm Supports: Evidence-Based Benefits and Criteria for Use , 2008, The journal of spinal cord medicine.

[11]  Thomas Bock,et al.  Intuitive and Adaptive Robotic Arm Manipulation using the Leap Motion Controller , 2014, ISR 2014.

[12]  S. Garber,et al.  Upper extremity assistive devices: assessment of use by spinal cord-injured patients with quadriplegia. , 1990, The American journal of occupational therapy : official publication of the American Occupational Therapy Association.

[13]  Diane J. Cook,et al.  Ambient intelligence and wearable computing: Sensors on the body, in the home, and beyond , 2009, J. Ambient Intell. Smart Environ..

[14]  J. Wiener,et al.  Measuring the activities of daily living: comparisons across national surveys. , 1990, Journal of gerontology.

[15]  Gregory D. Abowd,et al.  The Aware Home: A Living Laboratory for Ubiquitous Computing Research , 1999, CoBuild.

[16]  Achim J. Lilienthal,et al.  Comparative Evaluation of Range Sensor Accuracy in Indoor Environments , 2011, ECMR.

[17]  T. Hasegawa,et al.  A structured environment with sensor networks for intelligent robots , 2008, 2008 IEEE Sensors.

[18]  Thomas Bock,et al.  Towards a vision controlled robotic home environment , 2014 .

[19]  Francois Routhier,et al.  Evaluation of the JACO robotic arm: Clinico-economic study for powered wheelchair users with upper-extremity disabilities , 2011, 2011 IEEE International Conference on Rehabilitation Robotics.

[20]  William Harwin,et al.  Devices for assisting manipulation: a summary of user task priorities , 1994 .

[21]  Kanad K. Biswas,et al.  Gesture recognition using Microsoft Kinect® , 2011, The 5th International Conference on Automation, Robotics and Applications.

[22]  Hugh F. Durrant-Whyte,et al.  A solution to the simultaneous localization and map building (SLAM) problem , 2001, IEEE Trans. Robotics Autom..