Know Yourself and Know Others: Efficient Common Representation Learning for Few-shot Cross-modal Retrieval
暂无分享,去创建一个
[1] Dezhong Peng,et al. Deep Supervised Cross-Modal Retrieval , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[3] Nikos Komodakis,et al. Dynamic Few-Shot Visual Learning Without Forgetting , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[4] Wu-Jun Li,et al. Deep Cross-Modal Hashing , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Pedro H. O. Pinheiro,et al. Adaptive Cross-Modal Few-Shot Learning , 2019, NeurIPS.
[6] Wei Liu,et al. Self-Supervised Adversarial Hashing Networks for Cross-Modal Retrieval , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[7] Hanjiang Lai,et al. Simultaneous feature learning and hash coding with deep neural networks , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Hongxun Yao,et al. Hierarchical semantic image matching using CNN feature pyramid , 2018, Comput. Vis. Image Underst..
[9] Bharath Hariharan,et al. Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Xi Zhang,et al. Attention-Aware Deep Adversarial Hashing for Cross-Modal Retrieval , 2017, ECCV.
[11] Yang Yang,et al. Adversarial Cross-Modal Retrieval , 2017, ACM Multimedia.
[12] Chao Zhang,et al. Deep Joint-Semantics Reconstructing Hashing for Large-Scale Unsupervised Cross-Modal Retrieval , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[13] Hugo Jair Escalante,et al. The segmented and annotated IAPR TC-12 benchmark , 2010, Comput. Vis. Image Underst..
[14] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[15] Huimin Lu,et al. Unsupervised cross-modal retrieval through adversarial learning , 2017, 2017 IEEE International Conference on Multimedia and Expo (ICME).
[16] Huimin Lu,et al. Deep adversarial metric learning for cross-modal retrieval , 2019, World Wide Web.
[17] Tat-Seng Chua,et al. NUS-WIDE: a real-world web image database from National University of Singapore , 2009, CIVR '09.
[18] Wei Liu,et al. Pairwise Relationship Guided Deep Hashing for Cross-Modal Retrieval , 2017, AAAI.
[19] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[20] Mark J. Huiskes,et al. The MIR flickr retrieval evaluation , 2008, MIR '08.
[21] Guiguang Ding,et al. Collective Matrix Factorization Hashing for Multimodal Data , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[22] Nikos Paragios,et al. Data fusion through cross-modality metric learning using similarity-sensitive hashing , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[23] Roger Levy,et al. On the Role of Correlation and Abstraction in Cross-Modal Multimedia Retrieval , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[24] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[25] Jieping Ye,et al. A least squares formulation for canonical correlation analysis , 2008, ICML '08.
[26] Yizhou Wang,et al. Quantized Correlation Hashing for Fast Cross-Modal Search , 2015, IJCAI.
[27] Yu-Chiang Frank Wang,et al. A Closer Look at Few-shot Classification , 2019, ICLR.
[28] Patrick Jähnichen,et al. Cross-modal Hallucination for Few-shot Fine-grained Recognition , 2018, ArXiv.
[29] Yuan-Fang Wang,et al. Reinforced Cross-Modal Matching and Self-Supervised Imitation Learning for Vision-Language Navigation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Abhinav Gupta,et al. Non-local Neural Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[31] Joshua B. Tenenbaum,et al. Infinite Mixture Prototypes for Few-Shot Learning , 2019, ICML.
[32] Jun Wang,et al. Ranking-based Deep Cross-modal Hashing , 2019, AAAI.
[33] Philip S. Yu,et al. Deep Visual-Semantic Hashing for Cross-Modal Retrieval , 2016, KDD.
[34] Danqi Chen,et al. Reasoning With Neural Tensor Networks for Knowledge Base Completion , 2013, NIPS.
[35] Tao Xiang,et al. Learning to Compare: Relation Network for Few-Shot Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[36] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[37] Jungong Han,et al. Cross-View Retrieval via Probability-Based Semantics-Preserving Hashing , 2017, IEEE Transactions on Cybernetics.
[38] Jianmin Wang,et al. Semantics-preserving hashing for cross-view retrieval , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Dongqing Zhang,et al. Large-Scale Supervised Multimodal Hashing with Semantic Correlation Maximization , 2014, AAAI.
[40] Yan Huang,et al. ACMM: Aligned Cross-Modal Memory for Few-Shot Image and Sentence Matching , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[41] Kilian Q. Weinberger,et al. Distance Metric Learning for Large Margin Nearest Neighbor Classification , 2005, NIPS.
[42] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[43] Gregory R. Koch,et al. Siamese Neural Networks for One-Shot Image Recognition , 2015 .