HLVU: A New Challenge to Test Deep Understanding of Movies the Way Humans do
暂无分享,去创建一个
George Awad | Ian Soboroff | Keith Curtis | Shahzad Rajput | G. Awad | I. Soboroff | Keith Curtis | Shahzad Rajput
[1] Sanja Fidler,et al. MovieQA: Understanding Stories in Movies through Question-Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Jonathan G. Fiscus,et al. TRECVID 2019: An evaluation campaign to benchmark Video Activity Detection, Video Captioning and Matching, and Video Search & retrieval , 2019, TRECVID.
[3] Vicente Ordonez,et al. Moviescope: Large-scale Analysis of Movies using Multiple Modalities , 2019, ArXiv.
[4] Ellen M. Voorhees,et al. The TREC-8 Question Answering Track Report , 1999, TREC.
[5] Wang Wei,et al. Expressing Multimedia Content Using Semantics — A Vision , 2018, 2018 IEEE 12th International Conference on Semantic Computing (ICSC).
[6] Florian Metze,et al. How2: A Large-scale Dataset for Multimodal Language Understanding , 2018, NIPS 2018.
[7] Bernard Ghanem,et al. ActivityNet: A large-scale video benchmark for human activity understanding , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Margaret Mitchell,et al. VQA: Visual Question Answering , 2015, International Journal of Computer Vision.
[9] Alexander Dekhtyar,et al. Information Retrieval , 2018, Lecture Notes in Computer Science.
[10] Deva Ramanan,et al. Are we Asking the Right Questions in MovieQA? , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).