Hotspots detection for machine operation in egocentric vision

This paper introduces a novel idea of unsupervised hotspots detection from first person vision (FPV) records. The purpose is to gather typical patterns of machine operations based on touching or manipulating those hotspots and summarize the patterns as guides for operations such as online operating manuals. We chose sewing machine operation as an example and demonstrated that, a good performance of hotspots detection can be achieved by utilizing multiple features, especially touch and hand motion. More importantly, detected hotspots in both temporal and spatial locations matches well the positions of key components such as buttons, levers, and other important portions essential for operating the machine.

[1]  Dima Damen,et al.  You-Do, I-Learn: Discovering Task Relevant Objects and their Modes of Interaction from Multi-User Egocentric Video , 2014, BMVC.

[2]  Andrew D. Wilson Using a depth camera as a touch sensor , 2010, ITS '10.

[3]  Andrew Blake,et al.  High Precision Multi-touch Sensing on Surfaces using Overhead Cameras , 2007, Second Annual IEEE International Workshop on Horizontal Interactive Human-Computer Systems (TABLETOP'07).

[4]  Eric C. Larson,et al.  Dante vision: In-air and touch gesture sensing for natural surface interaction with combined depth and thermal cameras , 2012, 2012 IEEE International Conference on Emerging Signal Processing Applications.

[5]  Dima Damen,et al.  Estimating visual attention from a head mounted IMU , 2015, SEMWEB.

[6]  Deva Ramanan,et al.  Understanding Everyday Hands in Action from RGB-D Images , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[7]  Yoshiaki Shirai,et al.  Real-time 3D hand posture estimation based on 2D appearance retrieval using monocular camera , 2001, Proceedings IEEE ICCV Workshop on Recognition, Analysis, and Tracking of Faces and Gestures in Real-Time Systems.

[8]  Dieter Fox,et al.  Fine-grained kitchen activity recognition using RGB-D , 2012, UbiComp.