Research Article

Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges

Table 6

Evaluation measures of several abstractive text summarisation methods over the CNN/Daily Mail datasets.

ReferenceYearAuthorsModelROUGE1ROUGE2ROUGE-L

[55]2016Nallapati et al.Words-lvt2k-temp-att35.4613.3032.65
[56]2017See et al.Pointer-generator + coverage39.5317.2836.38
[57]2017Paulus et al.Reinforcement learning, with intra-attention41.1615.7539.08
[57]2017Paulus et al.Maximum-likelihood + RL, with intra-attention39.8715.8236.90
[58]2018Liu et al.Adversarial network39.9217.6536.71
[30]2018Song et al.ATSDL34.917.8
[35]2018Al-Sabahi et al.Bidirectional attentional encoder-decoder42.618.838.5
[59]2018Li et al.Key information guide network38.9517.1235.68
[60]2018Kryściński et al.ML + RL ROUGE + Novel, with LM40.1917.3837.52
[61]2018Yao et al.DEATS40.8518.0837.13
[62]2018Wan et al.BiSum37.0115.9533.66
[63]2019Wang et al.BEAR (large + WordPiece)41.9520.2639.49
[64]2019Egonmwan et al.TRANS-ext + filter + abs41.8918.938.92
[65]2020Liu et al.BERTSUMEXT (large)43.8520.3439.90
[49]2020Peng et al.DAPT + imp-coverage (RL + MLE (ss))40.7218.2837.35