X-BERT: eXtreme Multi-label Text Classification with BERT
暂无分享,去创建一个
Wei-Cheng Chang | Yiming Yang | Hsiang-Fu Yu | Kai Zhong | Inderjit Dhillon | I. Dhillon | Kai Zhong | Yiming Yang | Hsiang-Fu Yu | Wei-Cheng Chang
[1] Yiming Yang,et al. Deep Learning for Extreme Multi-label Text Classification , 2017, SIGIR.
[2] Pradeep Ravikumar,et al. PPDsparse: A Parallel Primal-Dual Sparse Method for Extreme Classification , 2017, KDD.
[3] Ali Mousavi,et al. Breaking the Glass Ceiling for Embedding-Based Classifiers for Large Output Spaces , 2019, NeurIPS.
[4] Ming-Wei Chang,et al. Latent Retrieval for Weakly Supervised Open Domain Question Answering , 2019, ACL.
[5] Jia Li,et al. Latent Cross: Making Use of Context in Recurrent Recommender Systems , 2018, WSDM.
[6] Pradeep Ravikumar,et al. PD-Sparse : A Primal and Dual Sparse Approach to Extreme Multiclass and Multilabel Classification , 2016, ICML.
[7] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[8] Zihan Zhang,et al. AttentionXML: Label Tree-based Attention-Aware Deep Model for High-Performance Extreme Multi-Label Text Classification , 2019, NeurIPS.
[9] Alex Wang,et al. What do you learn from context? Probing for sentence structure in contextualized word representations , 2019, ICLR.
[10] Pradeep Ravikumar,et al. Loss Decomposition for Fast Learning in Large Output Spaces , 2018, ICML.
[11] Hsuan-Tien Lin,et al. Feature-aware Label Space Dimension Reduction for Multi-label Classification , 2012, NIPS.
[12] Chih-Jen Lin,et al. LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..
[13] Rohit Babbar,et al. Bonsai - Diverse and Shallow Trees for Extreme Multi-label Classification , 2019, ArXiv.
[14] Manik Varma,et al. Extreme Multi-label Learning with Label Features for Warm-start Tagging, Ranking & Recommendation , 2018, WSDM.
[15] Moustapha Cissé,et al. Robust Bloom Filters for Large MultiLabel Classification Tasks , 2013, NIPS.
[16] 知秀 柴田. 5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding , 2020 .
[17] Manik Varma,et al. FastXML: a fast, accurate and stable tree-classifier for extreme multi-label learning , 2014, KDD.
[18] Wei-Cheng Chang,et al. Pre-training Tasks for Embedding-based Large-scale Retrieval , 2020, ICLR.
[19] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[20] Pasi Fränti,et al. Balanced K-Means for Clustering , 2014, S+SSPR.
[21] Sashank J. Reddi,et al. Stochastic Negative Mining for Learning with Large Output Spaces , 2018, AISTATS.
[22] Jürgen Schmidhuber,et al. Framewise phoneme classification with bidirectional LSTM and other neural network architectures , 2005, Neural Networks.
[23] Venkatesh Balasubramanian,et al. Slice: Scalable Linear Extreme Classifiers Trained on 100 Million Labels for Related Searches , 2019, WSDM.
[24] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[25] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[26] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[27] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[28] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[29] Jimmy J. Lin,et al. DocBERT: BERT for Document Classification , 2019, ArXiv.
[30] John Langford,et al. Multi-Label Prediction via Compressed Sensing , 2009, NIPS.
[31] Bernhard Schölkopf,et al. DiSMEC: Distributed Sparse Machines for Extreme Multi-label Classification , 2016, WSDM.
[32] Jason Weston,et al. Label Embedding Trees for Large Multi-Class Tasks , 2010, NIPS.
[33] Eyke Hüllermeier,et al. Extreme F-measure Maximization using Sparse Probability Estimates , 2016, ICML.
[34] Tong Zhang,et al. Supervised and Semi-Supervised Text Categorization using LSTM for Region Embeddings , 2016, ICML.
[35] Róbert Busa-Fekete,et al. A no-regret generalization of hierarchical softmax to extreme multi-label classification , 2018, NeurIPS.
[36] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[37] Ehsan Abbasnejad,et al. Label Filters for Large Scale Multilabel Classification , 2017, AISTATS.
[38] Johannes Fürnkranz,et al. Maximizing Subset Accuracy with Recurrent Neural Networks in Multi-label Classification , 2017, NIPS.
[39] Inderjit S. Dhillon,et al. Gradient Boosted Decision Trees for High Dimensional Sparse Output , 2017, ICML.
[40] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[41] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[42] Kyunghyun Cho,et al. Passage Re-ranking with BERT , 2019, ArXiv.
[43] Yukihiro Tagami,et al. AnnexML: Approximate Nearest Neighbor Search for Extreme Multi-label Classification , 2017, KDD.
[44] Inderjit S. Dhillon,et al. Large-scale Multi-label Learning with Missing Labels , 2013, ICML.
[45] Bernhard Schölkopf,et al. Data scarcity, robustness and extreme multi-label classification , 2019, Machine Learning.
[46] Tomas Mikolov,et al. Bag of Tricks for Efficient Text Classification , 2016, EACL.
[47] Prateek Jain,et al. Sparse Local Embeddings for Extreme Multi-label Classification , 2015, NIPS.
[48] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[49] Jason Weston,et al. WSABIE: Scaling Up to Large Vocabulary Image Annotation , 2011, IJCAI.
[50] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[51] Grigorios Tsoumakas,et al. Effective and Efficient Multilabel Classification in Domains with Large Number of Labels , 2008 .
[52] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[53] Yann Dauphin,et al. Pay Less Attention with Lightweight and Dynamic Convolutions , 2019, ICLR.
[54] Yann Dauphin,et al. Language Modeling with Gated Convolutional Networks , 2016, ICML.
[55] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[56] Manik Varma,et al. Multi-label learning with millions of labels: recommending advertiser bid phrases for web pages , 2013, WWW.
[57] Georgios Paliouras,et al. LSHTC: A Benchmark for Large-Scale Text Classification , 2015, ArXiv.
[58] Hiroshi Mamitsuka,et al. AttentionXML: Extreme Multi-Label Text Classification with Multi-Label Attention Based Recurrent Neural Networks , 2018, ArXiv.
[59] Shanfeng Zhu,et al. HAXMLNet: Hierarchical Attention Network for Extreme Multi-Label Text Classification , 2019, ArXiv.
[60] Jon Louis Bentley,et al. Multidimensional binary search trees used for associative searching , 1975, CACM.
[61] Jason Weston,et al. Label Partitioning For Sublinear Ranking , 2013, ICML.
[62] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[63] Manik Varma,et al. Extreme Multi-label Loss Functions for Recommendation, Tagging, Ranking & Other Missing Label Applications , 2016, KDD.
[64] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.