Command Palette
Search for a command to run...
PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization
Wen Xiao Iz Beltagy Giuseppe Carenini Arman Cohan

Abstract
We introduce PRIMERA, a pre-trained model for multi-document representation with a focus on summarization that reduces the need for dataset-specific architectures and large amounts of fine-tuning labeled data. PRIMERA uses our newly proposed pre-training objective designed to teach the model to connect and aggregate information across documents. It also uses efficient encoder-decoder transformers to simplify the processing of concatenated input documents. With extensive experiments on 6 multi-document summarization datasets from 3 different domains on zero-shot, few-shot and full-supervised settings, PRIMERA outperforms current state-of-the-art dataset-specific and pre-trained models on most of these settings with large margins. The code and pre-trained models can be found at \url{https://github.com/allenai/PRIMER}.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| multi-document-summarization-on-multi-news | PRIMER | ROUGE-1: 49.9 ROUGE-2: 21.1 ROUGE-L: 25.9 |
| multi-document-summarization-on-wcep | PRIMER | ROUGE-1: 46.1 ROUGE-2: 25.2 ROUGE-L: 37.9 |
| text-summarization-on-arxiv-summarization | PRIMER | ROUGE-1: 47.6 ROUGE-2: 20.8 ROUGE-L: 42.6 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.