An approach for behavior discovery using clustering of dynamics

As robots enter more complex application domains and start to interact autonomously with their surroundings and with humans, it becomes essential that they can efficiently represent and interpret their streams of sensor data and model the behavior of objects in their environment. To do this automatically and without the need for extensive a priori models of the environment, these systems have to be able to autonomously discover the different dynamic behaviors of entities in their environment as well as to predict points at which objects' behaviors and interactions change. This paper presents a technique to simultaneously learn to identify segmentation points and to extract a discrete set of models of the behaviors of objects from a stream of sensor observations. The approach presented here uses unsupervised learning techniques to learn these models in the form of representative sensor signatures which, once acquired, are used to interpret the robot's observations and to represent the actual sensor data more compactly as a sequence of behaviors. To enable the system to simultaneously learn to segment the continuous data stream and to build appropriate models for the data segments, a set of similarity metrics for sensor streams is derived and used in an expectation maximization algorithm. This algorithm alternates between segmenting the sensor data based on the existing behavior models and clustering the data segments to derive better models, thus iteratively improving both the quality of the segmentation and of the behavior models. To illustrate the approach, experiments are performed using simulated observations derived from different types of sensors which observe the dynamic interactions of objects.

[1]  Benjamin Kuipers,et al.  A robot exploration and mapping strategy based on a semantic hierarchy of spatial representations , 1991, Robotics Auton. Syst..

[2]  Jorma Rissanen,et al.  The Minimum Description Length Principle in Coding and Modeling , 1998, IEEE Trans. Inf. Theory.

[3]  Peter Bakker,et al.  Robot see, robot do: An overview of robot imitation , 1996 .

[4]  Tian Zhang,et al.  BIRCH: an efficient data clustering method for very large databases , 1996, SIGMOD '96.

[5]  Ashwin Ram,et al.  Continuous Case-Based Reasoning , 1997, Artif. Intell..

[6]  Claude E. Shannon,et al.  The Mathematical Theory of Communication , 1950 .

[7]  Phillip J. McKerrow,et al.  Introduction to robotics , 1991 .

[8]  Benjamin Kuipers,et al.  Map Learning with Uninterpreted Sensors and Effectors , 1995, Artif. Intell..

[9]  Mark A. Pitt,et al.  Advances in Minimum Description Length: Theory and Applications , 2005 .

[10]  Richard A. Johnson,et al.  Applied Multivariate Statistical Analysis , 1983 .

[11]  Kevin Barraclough,et al.  I and i , 2001, BMJ : British Medical Journal.

[12]  Mokshay Madiman,et al.  Minimum description length vs. maximum likelihood in lossy data compression , 2004, International Symposium onInformation Theory, 2004. ISIT 2004. Proceedings..

[13]  Jiawei Han,et al.  Data Mining: Concepts and Techniques , 2000 .

[14]  John Juyang Weng SHOSLIF: A Framework for Sensor-Based Learning for High-Dimensional Complex Systems , 1995 .

[15]  Paul S. Bradley,et al.  Clustering via Concave Minimization , 1996, NIPS.

[16]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[17]  Dearborn Animal Intelligence: An Experimental Study of the Associative Processes in Animals , 1900 .

[18]  James D. Murray Method of steepest descents , 1984 .

[19]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[20]  Ioannis Kontoyiannis,et al.  Maximum Likelihood Estimation for Lossy Data Compression , 2002 .

[21]  E. Oja,et al.  Independent Component Analysis , 2013 .

[22]  Laura Firoiu,et al.  Clustering Time Series with Hidden Markov Models and Dynamic Time Warping , 1999 .

[23]  Paul R. Cohen,et al.  Building a baby , 1996 .

[24]  Gregory Dudek,et al.  Learning environmental features for pose estimation , 2001, Image Vis. Comput..

[25]  J. Friedman Multivariate adaptive regression splines , 1990 .

[26]  Jean-Paul Laumond,et al.  Position referencing and consistent world modeling for mobile robots , 1985, Proceedings. 1985 IEEE International Conference on Robotics and Automation.

[27]  Claus Lamm,et al.  Single Trial Estimation of Evoked Potentials Using Gaussian Mixture Models with Integrated Noise Component , 2001, ICANN.

[28]  Spyridon Kamvissis On the Method of Steepest Descent , 2005 .

[29]  Sudipto Guha,et al.  CURE: an efficient clustering algorithm for large databases , 1998, SIGMOD '98.

[30]  Brian Everitt,et al.  Cluster analysis , 1974 .

[31]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[32]  Michael T. Rosenstein,et al.  Symbol Grounding With Delay Coordinates , 2003 .

[33]  Paul R. Cohen,et al.  Concepts From Time Series , 1998, AAAI/IAAI.

[34]  Jaideep Srivastava,et al.  Event detection from time series data , 1999, KDD '99.

[35]  Michael I. Jordan,et al.  Hierarchies of Adaptive Experts , 1991, NIPS.

[36]  Hamdy A. Taha,et al.  Operations Research an Introduction , 2007 .

[37]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[38]  Hugh Durrant-Whyte,et al.  Integration, coordination, and control of multi-sensor robot systems , 1987 .

[39]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[40]  Georg Dorffner,et al.  Growing Event Memories for Autonomous Robots , 2002 .

[41]  John J. Craig,et al.  Introduction to Robotics Mechanics and Control , 1986 .

[42]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[43]  Frank Dellaert,et al.  The Expectation Maximization Algorithm , 2002 .

[44]  Charles E. Heckler,et al.  Applied Multivariate Statistical Analysis , 2005, Technometrics.