Using BERT and Augmentation in Named Entity Recognition for Cybersecurity Domain
暂无分享,去创建一个
Natalia V. Loukachevitch | Mikhail Tikhomirov | N. Loukachevitch | Anastasiia Sirotina | Boris Dobrov | B. Dobrov | M. Tikhomirov | A. Sirotina
[1] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[2] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[3] Natalia V. Loukachevitch,et al. Named Entity Recognition in Information Security Domain for Russian , 2019, RANLP.
[4] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[5] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[6] Michael D. Iannacone,et al. Automatic Labeling for Entity Extraction in Cyber Security , 2013, ArXiv.
[7] Diyi Yang,et al. That’s So Annoying!!!: A Lexical and Frame-Semantic Embedding Based Data Augmentation Approach to Automatic Categorization of Annoying Behaviors using #petpeeve Tweets , 2015, EMNLP.
[8] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[9] Natalia V. Loukachevitch,et al. Combining Knowledge and CRF-Based Approach to Named Entity Recognition in Russian , 2016, AIST.
[10] Kai Zou,et al. EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks , 2019, EMNLP.
[11] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[12] Josef Steinberger,et al. The Second Cross-Lingual Challenge on Recognition, Normalization, Classification, and Linking of Named Entities across Slavic Languages , 2019 .
[13] Timothy W. Finin,et al. Extracting Cybersecurity Related Linked Data from Text , 2013, 2013 IEEE Seventh International Conference on Semantic Computing.
[14] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[15] Mikhail Arkhipov,et al. Adaptation of Deep Bidirectional Multilingual Transformers for Russian Language , 2019, ArXiv.
[16] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[17] Sosuke Kobayashi,et al. Contextual Augmentation: Data Augmentation by Words with Paradigmatic Relations , 2018, NAACL.