Learning Human Ergonomic Preferences for Handovers

Our goal is for people to be physically comfortable when taking objects from robots. This puts a burden on the robot to hand over the object in such a way that a person can easily reach it, without needing to strain or twist their arm - a way that is conducive to ergonomic human grasping configurations. To achieve this, the robot needs to understand what makes a configuration more or less ergonomic to the person, i.e. their ergonomic cost function. In this work, we formulate learning a person's ergonomic cost as an online estimation problem. The robot can implicitly make queries to the person by handing them objects in different configurations, and gets observations in response about the way they choose to take the object. We compare the performance of both passive and active approaches for solving this problem in simulation, as well as in an in-person user study.

[1]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[2]  Rachid Alami,et al.  Planning Safe and Legible Hand-over Motions for Human-Robot Interaction , 2010 .

[3]  Maya Cakmak,et al.  Designing Interactions for Robot Active Learners , 2010, IEEE Transactions on Autonomous Mental Development.

[4]  Rachid Alami,et al.  A Human-Aware Manipulation Planner , 2012, IEEE Transactions on Robotics.

[5]  David Maxwell Chickering,et al.  Here or there: preference judgments for relevance , 2008 .

[6]  Siddhartha S. Srinivasa,et al.  Task Space Regions , 2011, Int. J. Robotics Res..

[7]  Anca D. Dragan,et al.  Information gathering actions over human internal state , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Eyke Hüllermeier,et al.  Preference-based reinforcement learning: a formal framework and a policy iteration algorithm , 2012, Mach. Learn..

[9]  Subbarao Kambhampati,et al.  Probabilistic Planning via Determinization in Hindsight , 2008, AAAI.

[10]  S. Shankar Sastry,et al.  Personalized kinematics for human-robot collaborative manipulation , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[11]  Manjai Lee,et al.  Three Handover Methods in Esteem Etiquettes Using Dual Arms and Hands of Home-Service Robot , 2004 .

[12]  Nebojsa Jojic,et al.  Efficient Ranking from Pairwise Comparisons , 2013, ICML.

[13]  Eyke Hüllermeier,et al.  Label ranking by learning pairwise preferences , 2008, Artif. Intell..

[14]  Maya Cakmak,et al.  Adaptive Coordination Strategies for Human-Robot Handovers , 2015, Robotics: Science and Systems.

[15]  Rachid Alami,et al.  Sharing effort in planning human-robot handover tasks , 2012, 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication.

[16]  Robert Givan,et al.  A framework for simulation-based network control via hindsight optimization , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).

[17]  Craig Boutilier,et al.  A POMDP formulation of preference elicitation problems , 2002, AAAI/IAAI.

[18]  Martial Hebert,et al.  Active Learning For Outdoor Obstacle Detection , 2005, Robotics: Science and Systems.

[19]  Andrew McCallum,et al.  Active Learning by Labeling Features , 2009, EMNLP.

[20]  Manuel Lopes,et al.  Active Learning for Teaching a Robot Grounded Relational Symbols , 2013, IJCAI.

[21]  Artem Gritsenko,et al.  Learning Task-Specific Path-Quality Cost Functions From Expert Preferences , 2014 .

[22]  Amin Karbasi,et al.  Comparison-Based Learning with Rank Nets , 2012, ICML.

[23]  Anca D. Dragan,et al.  Active Preference-Based Learning of Reward Functions , 2017, Robotics: Science and Systems.

[24]  Oliver Kroemer,et al.  Active Reward Learning , 2014, Robotics: Science and Systems.

[25]  Ahmed Rashid It Takes Two Hands to Clap , 2005 .

[26]  Heni Ben Amor,et al.  Handover planning for every occasion , 2014, 2014 IEEE-RAS International Conference on Humanoid Robots.

[27]  Ali E. Abbas,et al.  Entropy methods for adaptive utility elicitation , 2004, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[28]  Eyal Amir,et al.  Bayesian Inverse Reinforcement Learning , 2007, IJCAI.

[29]  Siddhartha S. Srinivasa,et al.  Human preferences for robot-human hand-over configurations , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[30]  Siddhartha S. Srinivasa,et al.  Perception and Control Challenges for Effec tive Human-Robot Handoff s , 2011 .

[31]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[32]  Darius Braziunas,et al.  Computational Approaches to Preference Elicitation , 2006 .

[33]  Paul N. Bennett,et al.  Active Comparison Based Learning Incorporating User Uncertainty and Noise , 2016 .

[34]  Siddhartha S. Srinivasa,et al.  Toward seamless human-robot handovers , 2013, Journal of Human-Robot Interaction.

[35]  Anca D. Dragan,et al.  Implicitly Assisting Humans to Choose Good Grasps in Robot to Human Handovers , 2016, ISER.