Command Palette
Search for a command to run...
Unleashing the Power of Meta-tuning for Few-shot Generalization Through Sparse Interpolated Experts
Chen Shengzhuang ; Tack Jihoon ; Yang Yunqiao ; Teh Yee Whye ; Schwarz Jonathan Richard ; Wei Ying

Abstract
Recent successes suggest that parameter-efficient fine-tuning of foundationmodels as the state-of-the-art method for transfer learning in vision,replacing the rich literature of alternatives such as meta-learning. In tryingto harness the best of both worlds, meta-tuning introduces a subsequentoptimization stage of foundation models but has so far only shown limitedsuccess and crucially tends to underperform on out-of-distribution (OOD) tasks.In this paper, we introduce Sparse MetA-Tuning (SMAT), a method inspired bysparse mixture-of-experts approaches and trained to isolate subsets ofpre-trained parameters automatically for meta-tuning on each task. SMATsuccessfully overcomes OOD sensitivity and delivers on the promise of enhancingthe transfer abilities of vision foundation models beyond parameter-efficientfine-tuning. We establish new state-of-the-art results on a challengingcombination of Meta-Dataset augmented with additional OOD tasks in bothzero-shot and gradient-based adaptation settings. In addition, we provide athorough analysis of the superiority of learned over hand-designed sparsitypatterns for sparse expert methods and the pivotal importance of the sparsitylevel in balancing between in-distribution and out-of-distributiongeneralization. Our code is publicly available.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| few-shot-image-classification-on-meta-dataset | SMAT (DINO-VIT-Base-16-224) | Accuracy: 85.27 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.