Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2022 Aug 1:2022:7132226.
doi: 10.1155/2022/7132226. eCollection 2022.

A Comprehensive Survey of Abstractive Text Summarization Based on Deep Learning

Affiliations
Review

A Comprehensive Survey of Abstractive Text Summarization Based on Deep Learning

Mengli Zhang et al. Comput Intell Neurosci. .

Abstract

With the rapid development of the Internet, the massive amount of web textual data has grown exponentially, which has brought considerable challenges to downstream tasks, such as document management, text classification, and information retrieval. Automatic text summarization (ATS) is becoming an extremely important means to solve this problem. The core of ATS is to mine the gist of the original text and automatically generate a concise and readable summary. Recently, to better balance and develop these two aspects, deep learning (DL)-based abstractive summarization models have been developed. At present, for ATS tasks, almost all state-of-the-art (SOTA) models are based on DL architecture. However, a comprehensive literature survey is still lacking in the field of DL-based abstractive text summarization. To fill this gap, this paper provides researchers with a comprehensive survey of DL-based abstractive summarization. We first give an overview of abstractive summarization and DL. Then, we summarize several typical frameworks of abstractive summarization. After that, we also give a comparison of several popular datasets that are commonly used for training, validation, and testing. We further analyze the performance of several typical abstractive summarization systems on common datasets. Finally, we highlight some open challenges in the abstractive summarization task and outline some future research trends. We hope that these explorations will provide researchers with new insights into DL-based abstractive summarization.

PubMed Disclaimer

Conflict of interest statement

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Figures

Figure 1
Figure 1
A general architecture of DL-based ABS. It is mainly composed of three steps: preprocessing, semantic understanding, and summary generation.
Figure 2
Figure 2
RNN timeline expansion diagram.
Figure 3
Figure 3
Framework of convolutional neural networks.
Figure 4
Figure 4
The  schematic diagram of GNN. The  basic idea of GNN is to embed nodes according to the local neighbourhoods.
Figure 5
Figure 5
The basic encoder-decoder framework. It consists of input layer, hidden layer, and output layer.
Figure 6
Figure 6
The basic encoder-decoder framework with attention mechanisms. The attention mechanism enables the decoder to interact with the input during the decoding process.
Figure 7
Figure 7
The basic hierarchical encoder-decoder architecture. It is mainly divided into sentence level and word level. The word level processes each word token, and the sentence level processes each sentence.
Figure 8
Figure 8
The CNN-based ABS model. It is the most representative ABS model based entirely on CNN.
Figure 9
Figure 9
The framework of the pointer softmax. It utilizes two softmax layers to predict the next generated words: one softmax to predict the location of the word in the source sentence and copy it as output, and the other to predict the word in the shortlist vocabulary.
Figure 10
Figure 10
The framework of the PG model. It utilizes a pointer to copy words from the input document, which helps to accurately reproduce the information while retaining the ability to generate new tokens through the generator.
Figure 11
Figure 11
The overall framework of the FTSum model. It is a dual-attention encoder-decoder model.
Figure 12
Figure 12
The overall framework of the Entailment-aware encoder-decoder model. It uses the attention-based encoder-decoder framework as the infrastructure, and then shares the encoder with the entailment recognition system.
Figure 13
Figure 13
The overall framework of the FASum model. Its encoder and decoder are stacked by Transformer blocks.
Figure 14
Figure 14
The overall framework of the FAR-ASS model.

Similar articles

Cited by

References

    1. Khan A., Gul M. A., Zareei M., et al. Movie Review Summarization Using Supervised Learning and Graph-Basedranking Algorithm. Computational intelligence and neuroscience . 2020;2020 doi: 10.1155/2020/7526580.7526580 - DOI - PMC - PubMed
    1. Vilca G. C. V., Cabezudo M. A. S. A study of abstractive summarization using semantic representations and discourse level information. Text, Speech, and Dialogue . 2017:482–490. doi: 10.1007/978-3-319-64206-2_54. - DOI
    1. El-Kassas W. S., Salama C. R., Rafea A. A., Mohamed H. K. Automatic text summarization: a comprehensive survey. Expert Systems with Applications . 2021;165 doi: 10.1016/j.eswa.2020.113679.113679 - DOI
    1. Shen K., Hao P., Li R. A Compressive Sensing Model for Speeding up Text Classification. Computational Intelligence and Neuroscience . 2020 doi: 10.1155/2020/8879795.8879795 - DOI - PMC - PubMed
    1. Tao S., Shen C., Zhu L., Tao D. SVD-CNN: A Convolutional Neural Network Model with Orthogonal Constraints Based on SVD for Context-Aware Citation Recommendation. Computational Intelligence and Neuroscience . 2020 doi: 10.1155/2020/5343214.5343214 - DOI - PMC - PubMed

LinkOut - more resources