HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Multilingual Constituency Parsing with Self-Attention and Pre-Training

Nikita Kitaev; Steven Cao; Dan Klein

Multilingual Constituency Parsing with Self-Attention and Pre-Training

Abstract

We show that constituency parsing benefits from unsupervised pre-training across a variety of languages and a range of pre-training conditions. We first compare the benefits of no pre-training, fastText, ELMo, and BERT for English and find that BERT outperforms ELMo, in large part due to increased model capacity, whereas ELMo in turn outperforms the non-contextual fastText embeddings. We also find that pre-training is beneficial across all 11 languages tested; however, large model sizes (more than 100 million parameters) make it computationally expensive to train separate models for each language. To address this shortcoming, we show that joint multilingual pre-training and fine-tuning allows sharing all but a small number of parameters between ten languages in the final model. The 10x reduction in model size compared to fine-tuning one model per language causes only a 3.2% relative error increase in aggregate. We further explore the idea of joint fine-tuning and show that it gives low-resource languages a way to benefit from the larger datasets of other languages. Finally, we demonstrate new state-of-the-art results for 11 languages, including English (95.8 F1) and Chinese (91.8 F1).

Code Repositories

dpfried/rnng-bert
tf
Mentioned in GitHub
nikitakit/self-attentive-parser
Official
tf
Mentioned in GitHub
ghazalkhalighinejad/approximating-cky
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
constituency-parsing-on-ctb5Kitaev etal. 2019
F1 score: 91.75

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp