ESN-NER: entity storage network using attention mechanism for chinese NER

Chinese named entity recognition (NER) is more difficult than it in English because of the lack of nature delimiters. First, Chinese NER requires word segmentation, but word-based segmentation will generate errors due to the different granularity of the word segmentation tools. Second, most NER models heavily rely on local linguistic features, but the scope of influence provided by local linguistic features is limited, so sometimes the model will give different results to the same entity in different sentences. To address the above problems, we propose the Entity Storage Network Model called ESN Model for Chinese NER, which is a character-based model to avoid word segmentation errors. Specifically, we design an entity storage layer in this model to extract and store the entity information as a local linguistic feature, and design a position feature which is generated by four flags to enhance the learning of boundary. Then we incorporate the attention mechanism to extend the scope of the local linguistic features. The experimental results on two real-world datasets demonstrate that our model outperforms the state-of-the-art models in Chinese NER task.

[1]  Wei Xu,et al.  Bidirectional LSTM-CRF Models for Sequence Tagging , 2015, ArXiv.

[2]  Pavlina Fragkou,et al.  Applying named entity recognition and co-reference resolution for segmenting English texts , 2017, Progress in Artificial Intelligence.

[3]  Hideki Isozaki,et al.  Efficient Support Vector Classifiers for Named Entity Recognition , 2002, COLING.

[4]  Yue Zhang,et al.  Chinese NER Using Lattice LSTM , 2018, ACL.

[5]  Alex Graves,et al.  Neural Turing Machines , 2014, ArXiv.

[6]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[7]  Chandra Bhagavatula,et al.  Semi-supervised sequence tagging with bidirectional language models , 2017, ACL.

[8]  Marek Rei,et al.  Semi-supervised Multitask Learning for Sequence Labeling , 2017, ACL.

[9]  Ying Qin,et al.  Word Segmentation and Named Entity Recognition for SIGHAN Bakeoff3 , 2006, SIGHAN@COLING/ACL.

[10]  Richard M. Schwartz,et al.  Nymble: a High-Performance Learning Name-finder , 1997, ANLP.

[11]  Aitao Chen,et al.  Chinese Named Entity Recognition with Conditional Probabilistic Models , 2006, SIGHAN@COLING/ACL.

[12]  Yue Zhang,et al.  Multi-prototype Chinese Character Embedding , 2016, LREC.

[13]  Jun Zhao,et al.  Adversarial Transfer Learning for Chinese Named Entity Recognition with Self-Attention Mechanism , 2018, EMNLP.

[14]  James Hammerton,et al.  Named Entity Recognition with Long Short-Term Memory , 2003, CoNLL.

[15]  Dan Roth,et al.  Design Challenges and Misconceptions in Named Entity Recognition , 2009, CoNLL.

[16]  Guoxin Wang,et al.  CAN-NER: Convolutional Attention Network for Chinese Named Entity Recognition , 2019, NAACL.

[17]  Dan Roth,et al.  Entity Linking via Joint Encoding of Types, Descriptions, and Context , 2017, EMNLP.

[18]  Makoto Miwa,et al.  End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures , 2016, ACL.

[19]  Zaiqing Nie,et al.  Joint Entity Recognition and Disambiguation , 2015, EMNLP.

[20]  Ali Farhadi,et al.  Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.

[21]  Ruifang Liu,et al.  Neural Entity Reasoner for Global Consistency in NER , 2018, ArXiv.

[22]  Sampo Pyysalo,et al.  Attending to Characters in Neural Sequence Labeling Models , 2016, COLING.

[23]  Gina-Anne Levow,et al.  The Third International Chinese Language Processing Bakeoff: Word Segmentation and Named Entity Recognition , 2006, SIGHAN@COLING/ACL.

[24]  Masanori Hattori,et al.  Character-Based LSTM-CRF with Radical-Level Features for Chinese Named Entity Recognition , 2016, NLPCC/ICCPOL.

[25]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[26]  Hitoshi Isahara,et al.  Chinese Named Entity Recognition with Conditional Random Fields , 2006, SIGHAN@COLING/ACL.

[27]  Xiang Ren,et al.  Empower Sequence Labeling with Task-Aware Neural Language Model , 2017, AAAI.