Learning to navigate through crowded environments

The goal of this research is to enable mobile robots to navigate through crowded environments such as indoor shopping malls, airports, or downtown side walks. The key research question addressed in this paper is how to learn planners that generate human-like motion behavior. Our approach uses inverse reinforcement learning (IRL) to learn human-like navigation behavior based on example paths. Since robots have only limited sensing, we extend existing IRL methods to the case of partially observable environments. We demonstrate the capabilities of our approach using a realistic crowd flow simulator in which we modeled multiple scenarios in crowded environments. We show that our planner learned to guide the robot along the flow of people when the environment is crowded, and along the shortest path if no people are around.

[1]  Noel A Cressie,et al.  Statistics for Spatial Data. , 1992 .

[2]  Noel A Cressie,et al.  Statistics for Spatial Data, Revised Edition. , 1994 .

[3]  Andrew V. Goldberg,et al.  Strategic directions in research in theory of computing , 1996, CSUR.

[4]  Manfred K. Warmuth,et al.  Exponentiated Gradient Versus Gradient Descent for Linear Predictors , 1997, Inf. Comput..

[5]  Telecommunications Board Funding a Revolution: Government Support for Computing Research , 1999 .

[6]  Wolfram Burgard,et al.  Experiences with an Interactive Museum Tour-Guide Robot , 1999, Artif. Intell..

[7]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[8]  Paolo Fiorini,et al.  A robotics wheelchair for crowded public environment , 2001, IEEE Robotics Autom. Mag..

[9]  Paolo Fiorini,et al.  A Robotic Wheelchair for Crowded Public Environments MAid Combines Narrow- and Wide-Area Navigation Behaviors for Traversing Varying Environments While Retaining User Independence and Preventing Fatigue , 2001 .

[10]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[11]  Roland Siegwart,et al.  Robox at Expo.02: A large-scale installation of personal robots , 2003, Robotics Auton. Syst..

[12]  Reid G. Simmons,et al.  GRACE: An Autonomous Robot for the AAAI Robot Challenge , 2003, AI Mag..

[13]  Pieter Abbeel,et al.  Apprenticeship learning via inverse reinforcement learning , 2004, ICML.

[14]  Dieter Fox,et al.  Gaussian Processes for Signal Strength-Based Location Estimation , 2006, Robotics: Science and Systems.

[15]  J. Andrew Bagnell,et al.  Maximum margin planning , 2006, ICML.

[16]  Adrien Treuille,et al.  Continuum crowds , 2006, SIGGRAPH 2006.

[17]  Dani Lischinski,et al.  Crowds by Example , 2007, Comput. Graph. Forum.

[18]  Reid G. Simmons,et al.  Natural person-following behavior for social robots , 2007, 2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[19]  Dimitris N. Metaxas,et al.  Eurographics/ Acm Siggraph Symposium on Computer Animation (2007) Group Behavior from Video: a Data-driven Approach to Crowd Simulation , 2022 .

[20]  Reid G. Simmons,et al.  Socially Distributed Perception: GRACE plays social tag at AAAI 2005 , 2007, Auton. Robots.

[21]  Anind K. Dey,et al.  Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.

[22]  Anind K. Dey,et al.  Navigate like a cabbie: probabilistic reasoning from observed context-aware behavior , 2008, UbiComp.

[23]  Wolfram Burgard,et al.  Socially Inspired Motion Planning for Mobile Robots in Populated Environments , 2008 .

[24]  Sebastian Thrun,et al.  Apprenticeship learning for motion planning with application to parking lot navigation , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Dinesh Manocha,et al.  Interactive navigation of multiple agents in crowded environments , 2008, I3D '08.

[26]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.