HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Learning Bayesian Sparse Networks with Full Experience Replay for Continual Learning

Dong Gong; Qingsen Yan; Yuhang Liu; Anton van den Hengel; Javen Qinfeng Shi

Learning Bayesian Sparse Networks with Full Experience Replay for Continual Learning

Abstract

Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered. Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal. Despite their performance, they still suffer from interference across tasks which leads to catastrophic forgetting. To ameliorate this problem, we propose to only activate and select sparse neurons for learning current and past tasks at any stage. More parameters space and model capacity can thus be reserved for the future tasks. This minimizes the interference between parameters for different tasks. To do so, we propose a Sparse neural Network for Continual Learning (SNCL), which employs variational Bayesian sparsity priors on the activations of the neurons in all layers. Full Experience Replay (FER) provides effective supervision in learning the sparse activations of the neurons in different layers. A loss-aware reservoir-sampling strategy is developed to maintain the memory buffer. The proposed method is agnostic as to the network structures and the task boundaries. Experiments on different datasets show that our approach achieves state-of-the-art performance for mitigating forgetting.

Benchmarks

BenchmarkMethodologyMetrics
continual-learning-on-tiny-imagenet-10tasksA-GEM [chaudhry2018efficient]
Average Accuracy: 25.33
continual-learning-on-tiny-imagenet-10tasksDER [buzzega2020dark]
Average Accuracy: 51.78
continual-learning-on-tiny-imagenet-10tasksER[riemer2018learning]
Average Accuracy: 48.64
continual-learning-on-tiny-imagenet-10tasksSNCL
Average Accuracy: 52.85
continual-learning-on-tiny-imagenet-10tasksiCaRL [rebuffi2017icarl]
Average Accuracy: 31.55

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp