Deep Learning Transformer Architecture for Named-Entity Recognition on Low-Resourced Languages: State of the art results

This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named-Entity Recognition (NER) on ten low-resourced South African (SA) languages. In addition, these DL transformer models were com-pared to other Neural Network and Machine Learning (ML) NER models. The f...

Full description

Saved in:
Bibliographic Details
Published in2022 17th Conference on Computer Science and Intelligence Systems (FedCSIS) Vol. 30; pp. 53 - 60
Main Author Hanslo, Ridewaan
Format Conference Proceeding Journal Article
LanguageEnglish
Published Polish Information Processing Society 01.01.2022
Subjects
Online AccessGet full text
ISSN2300-5963
DOI10.15439/2022F53

Cover

Loading…
More Information
Summary:This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named-Entity Recognition (NER) on ten low-resourced South African (SA) languages. In addition, these DL transformer models were com-pared to other Neural Network and Machine Learning (ML) NER models. The findings show that transformer models substantially improve performance when applying discrete fine-tuning parameters per language. Furthermore, fine-tuned transformer models outperform other neural network and ma-chine learning models on NER with the low-resourced SA languages. For example, the transformer models obtained the highest F-scores for six of the ten SA languages and the highest average F-score surpassing the Conditional Random Fields ML model. Practical implications include developing high-performance NER capability with less effort and resource costs, potentially improving downstream NLP tasks such as Machine Translation (MT). Therefore, the application of DL trans-former architecture models for NLP NER sequence tagging tasks on low-resourced SA languages is viable. Additional re-search could evaluate the more recent transformer architecture models on other Natural Language Processing tasks and applications, such as Phrase chunking, MT, and Part-of-Speech tagging.
ISSN:2300-5963
DOI:10.15439/2022F53