BERT Representations for Video Question Answering
暂无分享,去创建一个
Chenhui Chu | Mayu Otani | Yuta Nakashima | Haruo Takemura | Noa Garcia | Zekun Yang | H. Takemura | Chenhui Chu | Mayu Otani | Yuta Nakashima | Zekun Yang | Noa García
[1] Matthieu Cord,et al. MUREL: Multimodal Relational Reasoning for Visual Question Answering , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Xing Gao,et al. Structured Semantic Representation for Visual Question Answering , 2018, 2018 25th IEEE International Conference on Image Processing (ICIP).
[3] Vicente Ordonez,et al. Obj2Text: Generating Visually Descriptive Language from Object Layouts , 2017, EMNLP.
[4] Christopher D. Manning,et al. Learning by Abstraction: The Neural State Machine , 2019, NeurIPS.
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[7] Licheng Yu,et al. TVQA: Localized, Compositional Video Question Answering , 2018, EMNLP.
[8] Ali Farhadi,et al. IQA: Visual Question Answering in Interactive Environments , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[9] Junyeong Kim,et al. Progressive Attention Memory Network for Movie Story Question Answering , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Yue-Shi Lee,et al. CLVQ: cross-language video question/answering system , 2004, IEEE Sixth International Symposium on Multimedia Software Engineering.
[11] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[12] Juan Carlos Niebles,et al. Leveraging Video Descriptions to Learn Video Question Answering , 2016, AAAI.
[13] Yu-Chieh Wu,et al. A Robust Passage Retrieval Algorithm for Video Question Answering , 2008, IEEE Transactions on Circuits and Systems for Video Technology.
[14] Jun Xiao,et al. Multi-Turn Video Question Answering via Multi-Stream Hierarchical Attention Context Network , 2018, IJCAI.
[15] Yu Long,et al. Static Correlative Filter Based Convolutional Neural Network for Visual Question Answering , 2018, 2018 IEEE International Conference on Big Data and Smart Computing (BigComp).
[16] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[17] Svetlana Lazebnik,et al. Out of the Box: Reasoning with Graph Convolution Nets for Factual Visual Question Answering , 2018, NeurIPS.
[18] Yi Yang,et al. Uncovering the Temporal Context for Video Question Answering , 2017, International Journal of Computer Vision.
[19] Jianyu Yang,et al. Multimodal Cross-guided Attention Networks for Visual Question Answering , 2018 .
[20] Wei Zhang,et al. R-VQA: Learning Visual Relation Facts with Semantic Attention for Visual Question Answering , 2018, KDD.
[21] Lei Zhang,et al. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[22] Ting Liu,et al. Story Ending Prediction by Transferable BERT , 2019, IJCAI.
[23] Sanja Fidler,et al. MovieQA: Understanding Stories in Movies through Question-Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[24] Tao Mei,et al. Tell-and-Answer: Towards Explainable Visual Question Answering using Attributes and Captions , 2018, EMNLP.
[25] Kate Saenko,et al. Ask, Attend and Answer: Exploring Question-Guided Spatial Attention for Visual Question Answering , 2015, ECCV.
[26] Bohyung Han,et al. MarioQA: Answering Questions by Watching Gameplay Videos , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[27] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[28] Di He,et al. Efficient Training of BERT by Progressively Stacking , 2019, ICML.
[29] Tomasz Kornuta,et al. Object-Based Reasoning in VQA , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).
[30] Juan Carlos Niebles,et al. Dense-Captioning Events in Videos , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[31] Byoung-Tak Zhang,et al. DeepStory: Video Story QA by Deep Embedded Memory Networks , 2017, IJCAI.
[32] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[33] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[34] Yash Goyal,et al. Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[35] Alexander G. Schwing,et al. Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answering , 2018, ECCV.
[36] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[37] Jiebo Luo,et al. VizWiz Grand Challenge: Answering Visual Questions from Blind People , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[38] Xinlei Chen,et al. Pythia-A platform for vision & language research , 2018 .
[39] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[40] Chenhui Chu,et al. KnowIT VQA: Answering Knowledge-Based Questions about Videos , 2020, AAAI.
[41] Yoshua Bengio,et al. Word Representations: A Simple and General Method for Semi-Supervised Learning , 2010, ACL.
[42] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[43] Byoung-Tak Zhang,et al. Bilinear Attention Networks , 2018, NeurIPS.
[44] Jianping Gou,et al. Mood-aware visual question answering , 2019, Neurocomputing.
[45] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[46] Licheng Yu,et al. TVQA+: Spatio-Temporal Grounding for Video Question Answering , 2019, ACL.
[47] Byoung-Tak Zhang,et al. Multimodal Dual Attention Memory for Video Story Question Answering , 2018, ECCV.