Extractive Text Summarization of Long Documents Using Word and Sentence Encoding

Automatic Text Summarization (ATS) for long documents is a very challenging task. A long document includes more than one topic, so it is required to construct a summary that covers the most important contextual information from the different topics in the input document. There are many ATS technique...

Full description

Saved in:
Bibliographic Details
Published in2023 5th Novel Intelligent and Leading Emerging Sciences Conference (NILES) pp. 130 - 135
Main Authors Abdelrahman, Amany A., El-Kassas, Wafaa S., Mohamed, Hoda K.
Format Conference Proceeding
LanguageEnglish
Published IEEE 21.10.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Automatic Text Summarization (ATS) for long documents is a very challenging task. A long document includes more than one topic, so it is required to construct a summary that covers the most important contextual information from the different topics in the input document. There are many ATS techniques, but the produced summaries are much less accurate than the human summaries. This paper proposes a supervised extractive Neural-based summarization model that focuses on long documents summarization by integrating both the local and global contexts of the document. The proposed system employs two bidirectional Gated Recurrent Units (GRUs) as word and sentence encoders, respectively, and uses Word2Vec embeddings as word representation. Two ensembles are proposed by applying the average ensemble method to the proposed system with two other Neural-based systems. The proposed system uses the PubMed dataset to evaluate the two ensembles using the ROUGE evaluation metrics. The results of the evaluation show that there are promising and remarkable improvements of 0.14% and 0.97% for ROUGE-1, 0.33% and 1.12% for ROUGE-2, and 0.25% for ROUGE-L.
DOI:10.1109/NILES59815.2023.10296690