暂无分享,去创建一个
Erkut Erdem | Aykut Erdem | Lucia Specia | Ozan Caglayan | Menekse Kuyu | Mustafa Sercan Amac | Pranava Madhyastha
[1] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[2] Khalil Sima'an,et al. Multi30K: Multilingual English-German Image Descriptions , 2016, VL@ACL.
[3] Jonathan Berant,et al. CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge , 2019, NAACL.
[4] Desmond Elliott,et al. DCU-UvA Multimodal MT System Report , 2016, WMT.
[5] Daniel Kondratyuk,et al. 75 Languages, 1 Model: Parsing Universal Dependencies Universally , 2019, EMNLP.
[6] Myle Ott,et al. Facebook FAIR’s WMT19 News Translation Task Submission , 2019, WMT.
[7] Jorma Laaksonen,et al. The MeMAD Submission to the WMT18 Multimodal Translation Task , 2018, WMT.
[8] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[9] Furu Wei,et al. MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers , 2020, NeurIPS.
[10] Svetlana Lazebnik,et al. Flickr30k Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[11] Joost van de Weijer,et al. Does Multimodality Help Human and Machine for Translation and Image Captioning? , 2016, WMT.
[12] Joost van de Weijer,et al. LIUM-CVC Submissions for WMT18 Multimodal Translation Task , 2018, WMT.
[13] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[14] Stefan Lee,et al. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks , 2019, NeurIPS.
[15] Desmond Elliott,et al. Imagination Improves Multimodal Translation , 2017, IJCNLP.
[16] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[17] Danqi Chen,et al. of the Association for Computational Linguistics: , 2001 .
[18] Peter Young,et al. From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions , 2014, TACL.
[19] Mohit Bansal,et al. LXMERT: Learning Cross-Modality Encoder Representations from Transformers , 2019, EMNLP.
[20] Sebastian Riedel,et al. Language Models as Knowledge Bases? , 2019, EMNLP.
[21] Mihika Dave. Multimodal machine translation , 2018 .
[22] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[23] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[24] Radu Soricut,et al. Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning , 2018, ACL.
[25] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[26] Nan Duan,et al. Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training , 2019, AAAI.
[27] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[28] Lucia Specia,et al. Probing the Need for Visual Context in Multimodal Machine Translation , 2019, NAACL.
[29] Jianfeng Gao,et al. Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks , 2020, ECCV.
[30] Jindrich Libovický,et al. CUNI System for the WMT18 Multimodal Translation Task , 2018, WMT.
[31] Jiebo Luo,et al. A Novel Graph-based Multi-modal Fusion Encoder for Neural Machine Translation , 2020, ACL.
[32] Yang Liu,et al. On Identifiability in Transformers , 2020, ICLR.
[33] Desmond Elliott,et al. Multimodal machine translation through visuals and speech , 2019, Machine Translation.
[34] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[35] Furu Wei,et al. VL-BERT: Pre-training of Generic Visual-Linguistic Representations , 2019, ICLR.
[36] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[37] Jindrich Libovický,et al. Attention Strategies for Multi-Source Sequence-to-Sequence Learning , 2017, ACL.
[38] Desmond Elliott,et al. Adversarial Evaluation of Multimodal Machine Translation , 2018, EMNLP.
[39] Stéphane Dupont,et al. UMONS Submission for WMT18 Multimodal Translation Task , 2018, ArXiv.
[40] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.