Toddler Sensory-Motor Development for Object Manipulation by Analyzing Hand-Pose

A toddler learns to hold various types of objects using hands in their everyday life without any specific instruction from an adult. The main media of interaction happens using the hands. A toddler's sensory-motor is developed by manipulating various types of objects. Exploring object holding-activity of a toddler can provide new insight into how sensory-motor develops in the formative years. Goal of this research is to learn how a toddler develops the sensory-motor system from directly hand interaction of an object. A video dataset was developed to study the patterns and hand-pose of a toddler while using their hands to hold an object. The birds-eye view of the scene was captured using a top-mounted camera, where the toddler was interacting with various types of objects with hands. Parents were guided to engage their child to interact with the object. The different hand-pose of the toddler's hands were observed, while the toddler was manipulating different types of objects. A detailed analysis result has presented on the hand-pose distribution of a toddler's hands for object holding-activities.

[1]  Andrew Zisserman,et al.  Hand detection using multiple proposals , 2011, BMVC.

[2]  Barbara Caputo,et al.  Recognizing human actions: a local SVM approach , 2004, ICPR 2004.

[3]  James M. Rehg,et al.  A Scalable Approach to Activity Recognition based on Object Use , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[4]  Barbara Caputo,et al.  Recognizing human actions: a local SVM approach , 2004, Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004..

[5]  Allan D. Jepson,et al.  Computational Perception of Scene Dynamics , 1996, ECCV.

[6]  James M. Rehg,et al.  Statistical Color Models with Application to Skin Detection , 2004, International Journal of Computer Vision.

[7]  King Ngi Ngan,et al.  Locating facial region of a head-and-shoulders color image , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[8]  Richard A. Foulds,et al.  Toward robust skin identification in video images , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[9]  Adil Mahmud Choudhury,et al.  A Machine Learning Approach to Identify Potential Customer Based on Purchase Behavior , 2019, 2019 International Conference on Robotics,Electrical and Signal Processing Techniques (ICREST).

[10]  Cordelia Schmid,et al.  Actions in context , 2009, CVPR.

[11]  Larry H. Matthies,et al.  First-Person Activity Recognition: What Are They Doing to Me? , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  RadhaKanta Mahapatra,et al.  Business data mining - a machine learning perspective , 2001, Inf. Manag..

[13]  Cheng Li,et al.  Model Recommendation with Virtual Probes for Egocentric Hand Detection , 2013, 2013 IEEE International Conference on Computer Vision.

[14]  Jake K. Aggarwal,et al.  Hierarchical Recognition of Human Activities Interacting with Objects , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[15]  S. Kolkur,et al.  Human Skin Detection Using RGB, HSV and YCbCr Color Models , 2017, Proceedings of the International Conference on Communication and Signal Processing 2016 (ICCASP 2016).

[16]  Yang Wang,et al.  Recognizing human actions from still images with latent poses , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[17]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[18]  Krupal S. Parikh,et al.  Support Vector Machine – A Large Margin Classifier to Diagnose Skin Illnesses , 2016 .

[19]  Deva Ramanan,et al.  Detecting activities of daily living in first-person camera views , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[20]  James M. Rehg,et al.  Learning to recognize objects in egocentric activities , 2011, CVPR 2011.

[21]  Cheng Li,et al.  Pixel-Level Hand Detection in Ego-centric Videos , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[22]  Irfan A. Essa,et al.  Exploiting human actions and object context for recognition tasks , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[23]  Larry S. Davis,et al.  Observing Human-Object Interactions: Using Spatial and Functional Compatibility for Recognition , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[24]  Abigail Sellen,et al.  Do life-logging technologies support memory for the past?: an experimental study using sensecam , 2007, CHI.

[25]  Adrian Hilton,et al.  A survey of advances in vision-based human motion capture and analysis , 2006, Comput. Vis. Image Underst..

[26]  Rama Chellappa,et al.  Machine Recognition of Human Activities: A Survey , 2008, IEEE Transactions on Circuits and Systems for Video Technology.

[27]  Steve E Hodges,et al.  Wearable cameras in health: the state of the art and future possibilities. , 2013, American journal of preventive medicine.

[28]  K. Marx,et al.  Applications of Machine Learning and High‐Dimensional Visualization in Cancer Detection, Diagnosis, and Management , 2004, Annals of the New York Academy of Sciences.

[29]  Fei-Fei Li,et al.  What, where and who? Classifying events by scene and object recognition , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[30]  E. Thelen Self organizing systems and infant motor development , 1987 .

[31]  James M. Rehg,et al.  Social interactions: A first-person perspective , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[32]  Stefan Lee,et al.  This Hand Is My Hand: A Probabilistic Approach to Hand Disambiguation in Egocentric Video , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[33]  Ali Farhadi,et al.  Understanding egocentric activities , 2011, 2011 International Conference on Computer Vision.

[34]  Xiaofeng Ren,et al.  Figure-ground segmentation improves handled object recognition in egocentric video , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[35]  Fei-Fei Li,et al.  Modeling mutual context of object and human pose in human-object interaction activities , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[36]  Cordelia Schmid,et al.  Actions in context , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[37]  David Murakami Wood,et al.  The Growth of CCTV: a global perspective on the international diffusion of video surveillance in publicly accessible space , 2002 .

[38]  Alexander H. Waibel,et al.  Skin-Color Modeling and Adaptation , 1998, ACCV.

[39]  Martin Paralic,et al.  Fast connected component labeling in binary images , 2012, 2012 35th International Conference on Telecommunications and Signal Processing (TSP).