학술논문

Towards Hybrid Model for Automatic Text Summarization
Document Type
Conference
Source
2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA) ICMLA Machine Learning and Applications (ICMLA), 2020 19th IEEE International Conference on. :987-993 Dec, 2020
Subject
Communication, Networking and Broadcast Technologies
Computing and Processing
Engineering Profession
Power, Energy and Industry Applications
Robotics and Control Systems
Signal Processing and Analysis
Recurrent neural networks
Semantics
Tools
Generators
Data models
Reliability
Testing
Extractive Summary
Recurrent Neural Network
Abstractive Summary
Pointer Generator Network
Language
Abstract
The overflowing of textual data on the web needs an efficient tool that is able to manage and process data. In this context, automatic text summarization has shown a great importance in several application areas. It aims to create a coherent and fluent short version of a document while preserving of the main information. This method allows for a reduction in reading time by condensing relevant information from a large collection of documents. Several automatic text summarization approaches have been proposed in order to entail shorten parts of the document. These methods have good results, but they still need improvements related to the reliability of sentences extraction, redundancy, semantic relationships between sentences, etc. This paper introduces a new hybrid architecture, combining a 2-layer recurrent neural network (RNN) extractive model and a sequence-to-sequence attentional abstractive model. This method uses the advantages of both extractive and abstractive approaches. A given text is first fed into the extractive model to obtain its relevant content, then generalized using the abstractive model, resulting in the text’s summary. First experimental results on real-world data show that the proposed model can achieve competitive results for extractive, abstractive and hybrid models.