Deep Learning Transformer Architecture for Named-Entity Recognition on Low-Resourced Languages: State of the art results
This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named-Entity Recognition (NER) on ten low-resourced South African (SA) languages. In addition, these DL transformer models were com-pared to other Neural Network and Machine Learning (ML) NER models. The f...
Saved in:
Published in | 2022 17th Conference on Computer Science and Intelligence Systems (FedCSIS) Vol. 30; pp. 53 - 60 |
---|---|
Main Author | |
Format | Conference Proceeding Journal Article |
Language | English |
Published |
Polish Information Processing Society
01.01.2022
|
Subjects | |
Online Access | Get full text |
ISSN | 2300-5963 |
DOI | 10.15439/2022F53 |
Cover
Loading…
Summary: | This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named-Entity Recognition (NER) on ten low-resourced South African (SA) languages. In addition, these DL transformer models were com-pared to other Neural Network and Machine Learning (ML) NER models. The findings show that transformer models substantially improve performance when applying discrete fine-tuning parameters per language. Furthermore, fine-tuned transformer models outperform other neural network and ma-chine learning models on NER with the low-resourced SA languages. For example, the transformer models obtained the highest F-scores for six of the ten SA languages and the highest average F-score surpassing the Conditional Random Fields ML model. Practical implications include developing high-performance NER capability with less effort and resource costs, potentially improving downstream NLP tasks such as Machine Translation (MT). Therefore, the application of DL trans-former architecture models for NLP NER sequence tagging tasks on low-resourced SA languages is viable. Additional re-search could evaluate the more recent transformer architecture models on other Natural Language Processing tasks and applications, such as Phrase chunking, MT, and Part-of-Speech tagging. |
---|---|
ISSN: | 2300-5963 |
DOI: | 10.15439/2022F53 |