HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

LAMOL: LAnguage MOdeling for Lifelong Language Learning

Fan-Keng Sun Cheng-Hao Ho Hung-Yi Lee

LAMOL: LAnguage MOdeling for Lifelong Language Learning

Abstract

Most research on lifelong learning applies to images or games, but not language. We present LAMOL, a simple yet effective method for lifelong language learning (LLL) based on language modeling. LAMOL replays pseudo-samples of previous tasks while requiring no extra memory or model capacity. Specifically, LAMOL is a language model that simultaneously learns to solve the tasks and generate training samples. When the model is trained for a new task, it generates pseudo-samples of previous tasks for training alongside data for the new task. The results show that LAMOL prevents catastrophic forgetting without any sign of intransigence and can perform five very different language tasks sequentially with only one model. Overall, LAMOL outperforms previous methods by a considerable margin and is only 2-3% worse than multitasking, which is usually considered the LLL upper bound. The source code is available at https://github.com/jojotenya/LAMOL.

Code Repositories

jojotenya/LAMOL
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
continual-learning-on-20newsgroup-10-tasksLAMOL
F1 - macro: 0.4572
continual-learning-on-asc-19-tasksLAMOL
F1 - macro: 0.8059

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp