暂无分享,去创建一个
[1] Vijay Mago,et al. TweetBERT: A Pretrained Language Representation Model for Twitter Text Analysis , 2020, ArXiv.
[2] Han Zhang,et al. Improving GANs Using Optimal Transport , 2018, ICLR.
[3] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[4] Xueqi Cheng,et al. Link Prediction in Knowledge Graphs: A Hierarchy-Constrained Approach , 2022, IEEE Transactions on Big Data.
[5] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[6] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[7] Jaewoo Kang,et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..
[8] Omer Levy,et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.
[9] Renjun Hu,et al. Self-Supervised Prototype Representation Learning for Event-Based Corporate Profiling , 2021, AAAI.
[10] Jianmo Ni,et al. Justifying Recommendations using Distantly-Labeled Reviews and Fine-Grained Aspects , 2019, EMNLP.
[11] Alessandro Rudi,et al. Differential Properties of Sinkhorn Approximation for Learning with Wasserstein Distance , 2018, NeurIPS.
[12] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[13] Yanchi Liu,et al. Job2Vec: Job Title Benchmarking with Collective Multi-View Representation Learning , 2019, CIKM.
[14] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[15] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[16] Xiaodong Liu,et al. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing , 2020, ACM Trans. Comput. Heal..
[17] Hui Xiong,et al. Market-oriented job skill valuation with cooperative composition neural network , 2021, Nature Communications.
[18] Yu Sun,et al. ERNIE: Enhanced Representation through Knowledge Integration , 2019, ArXiv.
[19] Bing Liu,et al. DomBERT: Domain-oriented Language Model for Aspect-based Sentiment Analysis , 2020, FINDINGS.
[20] Chao Zhang,et al. BOND: BERT-Assisted Open-Domain Named Entity Recognition with Distant Supervision , 2020, KDD.
[21] Hao Liu,et al. Spatio-Temporal Dual Graph Attention Network for Query-POI Matching , 2020, SIGIR.
[22] C. Villani. Optimal Transport: Old and New , 2008 .
[23] Iz Beltagy,et al. SciBERT: A Pretrained Language Model for Scientific Text , 2019, EMNLP.
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] Jiawei Han,et al. Automated Phrase Mining from Massive Text Corpora , 2017, IEEE Transactions on Knowledge and Data Engineering.
[26] Ramesh Nallapati,et al. Domain Adaptation with BERT-based Domain Classification and Data Selection , 2019, EMNLP.
[27] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[28] Manling Li,et al. Efficient parallel translating embedding for knowledge graphs , 2017, WI.
[29] Yu Cheng,et al. Graph Optimal Transport for Cross-Domain Alignment , 2020, ICML.
[30] Philip S. Yu,et al. BERT Post-Training for Review Reading Comprehension and Aspect-based Sentiment Analysis , 2019, NAACL.
[31] Haris Papageorgiou,et al. SemEval-2016 Task 5: Aspect Based Sentiment Analysis , 2016, *SEMEVAL.
[32] Sebastian Stabinger,et al. Adapt or Get Left Behind: Domain Adaptation through BERT Language Model Finetuning for Aspect-Target Sentiment Classification , 2020, LREC.
[33] Hao Tian,et al. ERNIE 2.0: A Continual Pre-training Framework for Language Understanding , 2019, AAAI.
[34] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[35] Hongyuan Zha,et al. A Fast Proximal Point Method for Computing Exact Wasserstein Distance , 2018, UAI.
[36] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[37] Marco Cuturi,et al. Sinkhorn Distances: Lightspeed Computation of Optimal Transport , 2013, NIPS.
[38] Ion Androutsopoulos,et al. LEGAL-BERT: “Preparing the Muppets for Court’” , 2020, FINDINGS.
[39] Benjamin Recht,et al. The Effect of Natural Distribution Shift on Question Answering Models , 2020, ICML.
[40] Jure Leskovec,et al. Inferring Networks of Substitutable and Complementary Products , 2015, KDD.
[41] Rajesh Ranganath,et al. ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission , 2019, ArXiv.