Sequence labeling model based on hierarchical features and attention mechanism

Sequence labeling is a basic task in natural language processing, which is of great help to processing text information. Conventional sequence labeling approaches heavily rely on hand-crafted or language-specific features, which requires a lot of time. Therefore, most of the existing methods are bas...

Full description

Saved in:
Bibliographic Details
Published inJournal of physics. Conference series Vol. 1883; no. 1; p. 12020
Main Authors Yao, Lijuan, Cheng, Yanfen, Li, Chao
Format Journal Article
LanguageEnglish
Published Bristol IOP Publishing 01.04.2021
Subjects
Online AccessGet full text
ISSN1742-6588
1742-6596
DOI10.1088/1742-6596/1883/1/012020

Cover

More Information
Summary:Sequence labeling is a basic task in natural language processing, which is of great help to processing text information. Conventional sequence labeling approaches heavily rely on hand-crafted or language-specific features, which requires a lot of time. Therefore, most of the existing methods are based on the BiLSTM-CRF model, but how to use a neural network to extract useful information for each unit or segment in the input sequence becomes the main factor limiting the efficiency. Several BiLSTM-CRF based models for sequence labeling have been presented, but the major limitation is how to use neural networks for extracting useful representations for each unit or segment in the input sequence. In response to this problem, this paper proposes a sequence labeling algorithm based on hierarchical features and attention mechanism, which uses a hierarchical structure to integrate character-level and word-level information, and applies different attention mechanisms to these two layers of information. According to the structural characteristics of different levels, excavate more potential information. Finally, the previously captured and guided features are used for sequence tag prediction using CRF. Finally, the proposed model is subjected to comparative experiments and the results obtained are analyzed.
Bibliography:ObjectType-Conference Proceeding-1
SourceType-Scholarly Journals-1
content type line 14
ISSN:1742-6588
1742-6596
DOI:10.1088/1742-6596/1883/1/012020