暂无分享,去创建一个
Nanyun Peng | Gully Burns | Sayan Paul | Te-Lin Wu | Shikhar Singh | Gully A. Burns | S. Paul | Nanyun Peng | Te-Lin Wu | Shikhar Singh
[1] Jingzhou Liu,et al. Violin: A Large-Scale Dataset for Video-and-Language Inference , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Ronald M. Summers,et al. Learning to Read Chest X-Rays: Recurrent Neural Cascade Model for Automated Image Annotation , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Haiyan Ding,et al. An Xception-GRU Model for Visual Question Answering in the Medical Domain , 2019, CLEF.
[4] Stefan Lee,et al. Embodied Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[5] Sanja Fidler,et al. MovieQA: Understanding Stories in Movies through Question-Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Mark Craven,et al. Constructing Biological Knowledge Bases by Extracting Information from Text Sources , 1999, ISMB.
[7] Lin Yang,et al. TandemNet: Distilling Knowledge from Medical Images Using Diagnostic Reports as Optional Semantic References , 2017, MICCAI.
[8] Francis Ferraro,et al. Visual Storytelling , 2016, NAACL.
[9] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[10] Jonghyun Choi,et al. Are You Smarter Than a Sixth Grader? Textbook Question Answering for Multimodal Machine Comprehension , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Aaron M. Cohen,et al. An Effective General Purpose Approach for Automated Biomedical Document Classification , 2006, AMIA.
[12] Georg Langs,et al. Predicting Semantic Descriptions from Medical Images with Convolutional Neural Networks , 2015, IPMI.
[13] Ali Farhadi,et al. A Diagram is Worth a Dozen Images , 2016, ECCV.
[14] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[15] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[16] Cho-Jui Hsieh,et al. VisualBERT: A Simple and Performant Baseline for Vision and Language , 2019, ArXiv.
[17] Ross B. Girshick,et al. Mask R-CNN , 2017, 1703.06870.
[18] Nanyun Peng,et al. Building deep learning models for evidence classification from the open access biomedical literature , 2019, Database J. Biol. Databases Curation.
[19] Yue Wu,et al. Towards Evidence Extraction : Analysis of Scientific Figures from Studies of Molecular Interactions , 2018, SemSci@ISWC.
[20] Ahmed El Kholy,et al. UNITER: Learning UNiversal Image-TExt Representations , 2019, ECCV 2020.
[21] Alexander J. Smola,et al. Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Ronald M. Summers,et al. TieNet: Text-Image Embedding Network for Common Thorax Disease Classification and Reporting in Chest X-Rays , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[23] Iz Beltagy,et al. SciBERT: A Pretrained Language Model for Scientific Text , 2019, EMNLP.
[24] Shizhe Chen,et al. YouMakeup: A Large-Scale Domain-Specific Multimodal Dataset for Fine-Grained Semantic Comprehension , 2019, EMNLP.
[25] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[26] Daniel Jurafsky,et al. Distant supervision for relation extraction without labeled data , 2009, ACL.
[27] Stefan Lee,et al. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks , 2019, NeurIPS.
[28] Hagit Shatkay,et al. Effective biomedical document classification for identifying publications relevant to the mouse Gene Expression Database (GXD) , 2017, Database J. Biol. Databases Curation.
[29] Licheng Yu,et al. TVQA: Localized, Compositional Video Question Answering , 2018, EMNLP.
[30] Nazli Ikizler-Cinbis,et al. RecipeQA: A Challenge Dataset for Multimodal Comprehension of Cooking Recipes , 2018, EMNLP.
[31] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[32] Zhiyong Lu,et al. A Fast Deep Learning Model for Textual Relevance in Biomedical Information Retrieval , 2018, WWW.
[33] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Zhiyong Lu,et al. Scaling up data curation using deep learning: An application to literature triage in genomic variation resources , 2018, PLoS Comput. Biol..
[35] Christian Simon,et al. BioReader: a text mining tool for performing classification of biomedical literature , 2019, BMC Bioinformatics.
[36] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[37] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[38] Eduard H. Hovy,et al. Extracting Evidence Fragments for Distant Supervision of Molecular Interactions , 2017, SemSci@ISWC.
[39] Judith A. Blake,et al. Integrating text mining into the MGI biocuration workflow , 2009, Database J. Biol. Databases Curation.
[40] Henning Müller,et al. VQA-Med: Overview of the Medical Visual Question Answering Task at ImageCLEF 2019 , 2019, CLEF.
[41] Pengtao Xie,et al. On the Automatic Generation of Medical Imaging Reports , 2017, ACL.
[42] Rafael C. Jimenez,et al. The MIntAct project—IntAct as a common curation platform for 11 molecular interaction databases , 2013, Nucleic Acids Res..
[43] Li Fei-Fei,et al. CLEVR: A Diagnostic Dataset for Compositional Language and Elementary Visual Reasoning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Thanh-Toan Do,et al. Overcoming Data Limitation in Medical Visual Question Answering , 2019, MICCAI.
[45] Michael S. Bernstein,et al. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations , 2016, International Journal of Computer Vision.
[46] K. Bretonnel Cohen,et al. Manual curation is not sufficient for annotation of genomic databases , 2007, ISMB/ECCB.
[47] Louis-Philippe Morency,et al. UR-FUNNY: A Multimodal Language Dataset for Understanding Humor , 2019, EMNLP.
[48] Abhishek Das,et al. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[49] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[50] Biocuration: Distilling data into knowledge , 2018, PLoS biology.
[51] Margaret Mitchell,et al. VQA: Visual Question Answering , 2015, International Journal of Computer Vision.
[52] Eric Xing,et al. PathVQA: 30000+ Questions for Medical Visual Question Answering , 2020, ArXiv.
[53] Ali Farhadi,et al. From Recognition to Cognition: Visual Commonsense Reasoning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Hagit Shatkay,et al. Integrating image data into biomedical text categorization , 2006, ISMB.
[55] Furu Wei,et al. VL-BERT: Pre-training of Generic Visual-Linguistic Representations , 2019, ICLR.
[56] George R. Thoma,et al. Design and Development of a Multimodal Biomedical Information Retrieval System , 2012, J. Comput. Sci. Eng..