HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training

UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training

Abstract

We propose to pre-train a unified language model for both autoencoding and partially autoregressive language modeling tasks using a novel training procedure, referred to as a pseudo-masked language model (PMLM). Given an input text with masked tokens, we rely on conventional masks to learn inter-relations between corrupted tokens and context via autoencoding, and pseudo masks to learn intra-relations between masked spans via partially autoregressive modeling. With well-designed position embeddings and self-attention masks, the context encodings are reused to avoid redundant computation. Moreover, conventional masks used for autoencoding provide global masking information, so that all the position embeddings are accessible in partially autoregressive language modeling. In addition, the two tasks pre-train a unified language model as a bidirectional encoder and a sequence-to-sequence decoder, respectively. Our experiments show that the unified language models pre-trained using PMLM achieve new state-of-the-art results on a wide range of natural language understanding and generation tasks across several widely used benchmarks.

Code Repositories

microsoft/dialoglm
pytorch
Mentioned in GitHub
facebookresearch/data2vec_vision
pytorch
Mentioned in GitHub
microsoft/unilm
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
abstractive-text-summarization-on-cnn-dailyUniLMv2
ROUGE-1: 43.16
ROUGE-2: 20.42
ROUGE-L: 40.14
question-generation-on-squad11UniLMv2
BLEU-4: 24.43

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp