HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization

Wen Xiao Iz Beltagy Giuseppe Carenini Arman Cohan

PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization

Abstract

We introduce PRIMERA, a pre-trained model for multi-document representation with a focus on summarization that reduces the need for dataset-specific architectures and large amounts of fine-tuning labeled data. PRIMERA uses our newly proposed pre-training objective designed to teach the model to connect and aggregate information across documents. It also uses efficient encoder-decoder transformers to simplify the processing of concatenated input documents. With extensive experiments on 6 multi-document summarization datasets from 3 different domains on zero-shot, few-shot and full-supervised settings, PRIMERA outperforms current state-of-the-art dataset-specific and pre-trained models on most of these settings with large margins. The code and pre-trained models can be found at \url{https://github.com/allenai/PRIMER}.

Code Repositories

allenai/open-mds
pytorch
Mentioned in GitHub
allenai/primer
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
multi-document-summarization-on-multi-newsPRIMER
ROUGE-1: 49.9
ROUGE-2: 21.1
ROUGE-L: 25.9
multi-document-summarization-on-wcepPRIMER
ROUGE-1: 46.1
ROUGE-2: 25.2
ROUGE-L: 37.9
text-summarization-on-arxiv-summarizationPRIMER
ROUGE-1: 47.6
ROUGE-2: 20.8
ROUGE-L: 42.6

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp