An Effective and Efficient Method for Detecting Hands in Egocentric Videos for Rehabilitation Applications

Individuals with spinal cord injury (SCI) report upper limb function as their top recovery priority. To accurately represent the true impact of new interventions on patient function, evaluation should occur in a natural setting. Wearable cameras can be used to monitor hand function at home, using computer vision to automatically analyze the resulting egocentric videos. A key step in this process, hand detection, is difficult to accomplish robustly and reliably, hindering the deployment of a complete monitoring system in the home and community. We propose an accurate and efficient hand detection method that uses a simple combination of existing detection and tracking algorithms, evaluated on a new hand detection dataset, consisting of 167,622 frames of egocentric videos collected from 17 individuals with SCI in a home simulation laboratory. The F1-scores for the best detector and tracker alone (SSD and Median Flow) were 0.90±0.07 and 0.42±0.18, respectively. The best combination method, in which a detector was used to initialize and reset a tracker, resulted in an F1-score of 0.87±0.07 while being two times faster than the fastest detector. The method proposed here, in combination with wearable cameras, will help clinicians directly measure hand function in a patient’s daily life at home.

[1]  Ali Farhadi,et al.  YOLO9000: Better, Faster, Stronger , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Cheng Li,et al.  Pixel-Level Hand Detection in Ego-centric Videos , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[3]  Aaron M. Dollar,et al.  Analyzing at-home prosthesis use in unilateral upper-limb amputees to inform treatment & device design , 2017, 2017 International Conference on Rehabilitation Robotics (ICORR).

[4]  C. V. Jawahar,et al.  First Person Action Recognition Using Deep Learned Descriptors , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  G. Scivoletto,et al.  A multicenter international study on the Spinal Cord Independence Measure, version III: Rasch psychometric validation , 2007, Spinal Cord.

[6]  H. Krueger,et al.  The economic burden of traumatic spinal cord injury in Canada. , 2013, Chronic diseases and injuries in Canada.

[7]  Stefan Lee,et al.  Lending A Hand: Detecting Hands and Recognizing Activities in Complex Egocentric Interactions , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[8]  W. Donovan,et al.  International Standards For Neurological Classification Of Spinal Cord Injury , 2003, The journal of spinal cord medicine.

[9]  Horst Bischof,et al.  Real-Time Tracking via On-line Boosting , 2006, BMVC.

[10]  David J. Reinkensmeyer,et al.  The Manumeter: A Wearable Device for Monitoring Daily Use of the Wrist and Fingers , 2014, IEEE Journal of Biomedical and Health Informatics.

[11]  K. Anderson Targeting recovery: priorities of the spinal cord-injured population. , 2004, Journal of neurotrauma.

[12]  H. Rodgers,et al.  Accelerometer measurement of upper extremity movement after stroke: a systematic review of clinical studies , 2014, Journal of NeuroEngineering and Rehabilitation.

[13]  Yingjie Cai,et al.  Multiple Object Tracking Based on Faster-RCNN Detector and KCF Tracker , 2016 .

[14]  Aaron M. Dollar,et al.  Grasp Frequency and Usage in Daily Household and Machine Shop Tasks , 2013, IEEE Transactions on Haptics.

[15]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[16]  Jirapat Likitlersuang,et al.  Egocentric video: a new tool for capturing hand use of individuals with spinal cord injury at home , 2019, Journal of NeuroEngineering and Rehabilitation.

[17]  Andrew Zisserman,et al.  Detect to Track and Track to Detect , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[18]  James M. Rehg,et al.  Learning to recognize objects in egocentric activities , 2011, CVPR 2011.

[19]  Christine E. King,et al.  Wearable Wireless Sensors for Rehabilitation , 2016 .

[20]  Ninja P. Oess,et al.  Design and evaluation of a low-cost instrumented glove for hand function assessment , 2012, Journal of NeuroEngineering and Rehabilitation.

[21]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[22]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  M. Granat,et al.  Upper limb activity of twenty myoelectric prosthesis users and twenty healthy anatomically intact adults , 2019, Scientific Data.

[24]  Andrei Krassioukov,et al.  International standards for neurological classification of spinal cord injury, revised 2011. , 2012, Topics in spinal cord injury rehabilitation.

[25]  Michifumi Yoshioka,et al.  Cooking activities recognition in egocentric videos using hand shape feature with openpose , 2018, CEA@IJCAI.

[26]  Matthias Rauterberg,et al.  A Dynamic Approach and a New Dataset for Hand-detection in First Person Vision , 2015, CAIP.

[27]  Frédéric Lerasle,et al.  A comparative view on exemplar ‘tracking-by-detection’ approaches , 2015, 2015 12th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS).

[29]  Stefan Roth,et al.  People-tracking-by-detection and people-detection-by-tracking , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Subhas Chandra Mukhopadhyay,et al.  Wearable Sensors for Human Activity Monitoring: A Review , 2015, IEEE Sensors Journal.

[31]  Ming-Hsuan Yang,et al.  Visual tracking with online Multiple Instance Learning , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[32]  Roger Gassert,et al.  Monitoring Upper Limb Recovery after Cervical Spinal Cord Injury: Insights beyond Assessment Scores , 2016, Front. Neurol..

[33]  Rui Caseiro,et al.  Exploiting the Circulant Structure of Tracking-by-Detection with Kernels , 2012, ECCV.

[34]  Kris M. Kitani,et al.  How do we use our hands? Discovering a diverse set of common grasps , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Paolo Bonato,et al.  The Use of a Finger-Worn Accelerometer for Monitoring of Hand Use in Ambulatory Settings , 2019, IEEE Journal of Biomedical and Health Informatics.

[36]  M. Popovic,et al.  The Graded Redefined Assessment of Strength Sensibility and Prehension: reliability and validity. , 2012, Journal of neurotrauma.

[37]  José Zariffa,et al.  Views of individuals with spinal cord injury on the use of wearable cameras to monitor upper limb function in the home and community , 2017, The journal of spinal cord medicine.

[38]  Jiri Matas,et al.  Forward-Backward Error: Automatic Detection of Tracking Failures , 2010, 2010 20th International Conference on Pattern Recognition.

[39]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[40]  Shyamal Patel,et al.  A review of wearable sensors and systems with application in rehabilitation , 2012, Journal of NeuroEngineering and Rehabilitation.

[41]  Jirapat Likitlersuang,et al.  Interaction Detection in Egocentric Video: Toward a Novel Outcome Measure for Upper Extremity Function , 2018, IEEE Journal of Biomedical and Health Informatics.

[42]  Horst Bischof,et al.  On-line Boosting and Vision , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[43]  Silvio Savarese,et al.  Learning to Track at 100 FPS with Deep Regression Networks , 2016, ECCV.

[44]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).