Co-training an Improved Recurrent Neural Network with Probability Statistic Models for Named Entity Recognition

Named Entity Recognition (NER) is a subtask of information extraction in Natural Language Processing (NLP) field and thus being wildly studied. Currently Recurrent Neural Network (RNN) has become a popular way to do NER task, but it needs a lot of train data. The lack of labeled train data is one of...

Full description

Saved in:
Bibliographic Details
Published inDatabase Systems for Advanced Applications Vol. 10178; pp. 545 - 555
Main Authors Sun, Yueqing, Li, Lin, Xie, Zhongwei, Xie, Qing, Li, Xin, Xu, Guandong
Format Book Chapter
LanguageEnglish
Published Switzerland Springer International Publishing AG 2017
Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text
ISBN3319556983
9783319556987
ISSN0302-9743
1611-3349
DOI10.1007/978-3-319-55699-4_33

Cover

Loading…
More Information
Summary:Named Entity Recognition (NER) is a subtask of information extraction in Natural Language Processing (NLP) field and thus being wildly studied. Currently Recurrent Neural Network (RNN) has become a popular way to do NER task, but it needs a lot of train data. The lack of labeled train data is one of the hard problems and traditional co-training strategy is a way to alleviate it. In this paper, we consider this situation and focus on doing NER with co-training using RNN and two probability statistic models i.e. Hidden Markov Model (HMM) and Conditional Random Field (CRF). We proposed a modified RNN model by redefining its activation function. Compared to traditional sigmoid function, our new function avoids saturation to some degree and makes its output scope very close to [0, 1], thus improving recognition accuracy. Our experiments are conducted ATIS benchmark. First, supervised learning using those models are compared when using different train data size. The experimental results show that it is not necessary to use whole data, even small part of train data can also get good performance. Then, we compare the results of our modified RNN with original RNN. 0.5% improvement is obtained. Last, we compare the co-training results. HMM and CRF get higher improvement than RNN after co-training. Moreover, using our modified RNN in co-training, their performances are improved further.
Bibliography:This research project is supported by the National Social Science Foundation of China (Grant No:15BGL048), National Natural Science Foundation of China (Grant No:61602353, 61303029), 863 Program (2015AA015403), Hubei Province Science and Technology Support Project (2015BAA072).
ISBN:3319556983
9783319556987
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-319-55699-4_33