Actively Seeking and Learning From Live Data
暂无分享,去创建一个
[1] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[2] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[3] Gaurav Sharma,et al. An Empirical Evaluation of Visual Question Answering for Novel Objects , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Trevor Darrell,et al. Multimodal Compact Bilinear Pooling for Visual Question Answering and Visual Grounding , 2016, EMNLP.
[5] Yuandong Tian,et al. Simple Baseline for Visual Question Answering , 2015, ArXiv.
[6] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[7] José M. F. Moura,et al. Visual Dialog , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[8] Trevor Darrell,et al. Adapting Visual Category Models to New Domains , 2010, ECCV.
[9] Hexiang Hu,et al. Cross-Dataset Adaptation for Visual Question Answering , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[10] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[11] Alexander J. Smola,et al. Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Lei Zhang,et al. Bottom-Up and Top-Down Attention for Image Captioning and VQA , 2017, ArXiv.
[13] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[14] Jannis Bulian,et al. Ask the Right Questions: Active Question Reformulation with Reinforcement Learning , 2017, ICLR.
[15] Anton van den Hengel,et al. Visual Question Answering as a Meta Learning Task , 2017, ECCV.
[16] Kyunghyun Cho,et al. Task-Oriented Query Reformulation with Reinforcement Learning , 2017, EMNLP.
[17] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[18] Sergey Levine,et al. One-Shot Visual Imitation Learning via Meta-Learning , 2017, CoRL.
[19] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[20] Chenxi Liu,et al. Deep Nets: What have They Ever Done for Vision? , 2018, International Journal of Computer Vision.
[21] Dhruv Batra,et al. C-VQA: A Compositional Split of the Visual Question Answering (VQA) v1.0 Dataset , 2017, ArXiv.
[22] Peng Wang,et al. Ask Me Anything: Free-Form Visual Question Answering Based on Knowledge from External Sources , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Chunhua Shen,et al. Explicit Knowledge-based Reasoning for Visual Question Answering , 2015, IJCAI.
[24] Sergey Levine,et al. One-Shot Imitation from Observing Humans via Domain-Adaptive Meta-Learning , 2018, Robotics: Science and Systems.
[25] Qi Wu,et al. Visual question answering: A survey of methods and datasets , 2016, Comput. Vis. Image Underst..
[26] Anton van den Hengel,et al. Tips and Tricks for Visual Question Answering: Learnings from the 2017 Challenge , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[27] Yoshua Bengio,et al. On the Optimization of a Synaptic Learning Rule , 2007 .
[28] Anton van den Hengel,et al. Zero-Shot Visual Question Answering , 2016, ArXiv.
[29] Richard J. Mammone,et al. Meta-neural networks that learn by learning , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.
[30] Michael S. Bernstein,et al. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations , 2016, International Journal of Computer Vision.
[31] Sebastian Thrun,et al. Lifelong Learning Algorithms , 1998, Learning to Learn.
[32] Regina Barzilay,et al. Improving Information Extraction by Acquiring External Evidence with Reinforcement Learning , 2016, EMNLP.
[33] Stefan Lee,et al. Overcoming Language Priors in Visual Question Answering with Adversarial Regularization , 2018, NeurIPS.
[34] Michael S. Bernstein,et al. Visual7W: Grounded Question Answering in Images , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[35] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[36] Lei Zhang,et al. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[37] Margaret Mitchell,et al. VQA: Visual Question Answering , 2015, International Journal of Computer Vision.
[38] Yoshua Bengio,et al. Learning a synaptic learning rule , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[39] Yash Goyal,et al. Yin and Yang: Balancing and Answering Binary Visual Questions , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[40] Yash Goyal,et al. Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[41] Dhruv Batra,et al. Don't Just Assume; Look and Answer: Overcoming Priors for Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[42] Xinlei Chen,et al. Microsoft COCO Captions: Data Collection and Evaluation Server , 2015, ArXiv.
[43] J. Urgen Schmidhuber. Learning to Control Fast-weight Memories: an Alternative to Dynamic Recurrent Networks , 1991 .
[44] Sebastian Thrun,et al. Learning to Learn , 1998, Springer US.
[45] Po-Sen Huang,et al. Natural Language to Structured Query Generation via Meta-Learning , 2018, NAACL.
[46] Gary Marcus,et al. Deep Learning: A Critical Appraisal , 2018, ArXiv.