Command Palette
Search for a command to run...
A Reinforced Topic-Aware Convolutional Sequence-to-Sequence Model for Abstractive Text Summarization
Li Wang; Junlin Yao; Yunzhe Tao; Li Zhong; Wei Liu; Qiang Du

Abstract
In this paper, we propose a deep learning approach to tackle the automatic summarization tasks by incorporating topic information into the convolutional sequence-to-sequence (ConvS2S) model and using self-critical sequence training (SCST) for optimization. Through jointly attending to topics and word-level alignment, our approach can improve coherence, diversity, and informativeness of generated summaries via a biased probability generation mechanism. On the other hand, reinforcement training, like SCST, directly optimizes the proposed model with respect to the non-differentiable metric ROUGE, which also avoids the exposure bias during inference. We carry out the experimental evaluation with state-of-the-art methods over the Gigaword, DUC-2004, and LCSTS datasets. The empirical results demonstrate the superiority of our proposed method in the abstractive summarization.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| text-summarization-on-duc-2004-task-1 | Reinforced-Topic-ConvS2S | ROUGE-1: 31.15 ROUGE-2: 10.85 ROUGE-L: 27.68 |
| text-summarization-on-gigaword | Reinforced-Topic-ConvS2S | ROUGE-1: 36.92 ROUGE-2: 18.29 ROUGE-L: 34.58 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.