Demo2Vec: Reasoning Object Affordances from Online Videos
暂无分享,去创建一个
Silvio Savarese | Joseph J. Lim | Kuan Fang | Te-Lin Wu | Daniel Yang | S. Savarese | Kuan Fang | Te-Lin Wu | Daniel Yang
[1] Graham W. Taylor,et al. Deconvolutional networks , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[2] Nikolaos G. Tsagarakis,et al. Detecting object affordances with Convolutional Neural Networks , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[3] Sinisa Todorovic,et al. A Multi-scale CNN for Affordance Segmentation in RGB Images , 2016, ECCV.
[4] Fei-Fei Li,et al. Discovering Object Functionality , 2013, 2013 IEEE International Conference on Computer Vision.
[5] Danica Kragic,et al. Visual object-action recognition: Inferring object affordances from human demonstration , 2011, Comput. Vis. Image Underst..
[6] Frédo Durand,et al. A Benchmark of Computational Models of Saliency to Predict Human Fixations , 2012 .
[7] Dit-Yan Yeung,et al. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting , 2015, NIPS.
[8] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[9] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[10] Siddhant Gangapurwala. Generative Adversarial Imitation Learning for Quadrupedal Locomotion using Unstructured Expert Demonstrations , 2018 .
[11] Pieter Abbeel,et al. Third-Person Imitation Learning , 2017, ICLR.
[12] Andrew Zisserman,et al. Spatial Transformer Networks , 2015, NIPS.
[13] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[14] Pieter Abbeel,et al. Learning from Demonstrations Through the Use of Non-rigid Registration , 2013, ISRR.
[15] Michael S. Ryoo,et al. Learning Social Affordance for Human-Robot Interaction , 2016, IJCAI.
[16] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[17] Hema Swetha Koppula,et al. Physically Grounded Spatio-temporal Object Affordances , 2014, ECCV.
[18] Juergen Gall,et al. Weakly Supervised Learning of Affordances , 2016, ArXiv.
[19] Arkanath Pathak,et al. Learning Grasping Interaction with Geometry-aware 3D Representations , 2017, ArXiv.
[20] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[21] Chenfanfu Jiang,et al. Inferring Forces and Learning Human Utilities from Videos , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Jürgen Schmidhuber,et al. Learning to forget: continual prediction with LSTM , 1999 .
[23] Jitendra Malik,et al. Combining self-supervised learning and imitation for vision-based rope manipulation , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[24] Frédo Durand,et al. What Do Different Evaluation Metrics Tell Us About Saliency Models? , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[25] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[26] Hema Swetha Koppula,et al. Learning human activities and object affordances from RGB-D videos , 2012, Int. J. Robotics Res..
[27] Marcin Andrychowicz,et al. One-Shot Imitation Learning , 2017, NIPS.
[28] Hema Swetha Koppula,et al. Anticipating Human Activities Using Object Affordances for Reactive Robotic Response , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[29] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[30] Song-Chun Zhu,et al. Understanding tools: Task-oriented object modeling, learning and recognition , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Yiannis Aloimonos,et al. Affordance detection of tool parts from geometric features , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[32] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.