HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Unleashing the Power of Meta-tuning for Few-shot Generalization Through Sparse Interpolated Experts

Chen Shengzhuang ; Tack Jihoon ; Yang Yunqiao ; Teh Yee Whye ; Schwarz Jonathan Richard ; Wei Ying

Unleashing the Power of Meta-tuning for Few-shot Generalization Through
  Sparse Interpolated Experts

Abstract

Recent successes suggest that parameter-efficient fine-tuning of foundationmodels as the state-of-the-art method for transfer learning in vision,replacing the rich literature of alternatives such as meta-learning. In tryingto harness the best of both worlds, meta-tuning introduces a subsequentoptimization stage of foundation models but has so far only shown limitedsuccess and crucially tends to underperform on out-of-distribution (OOD) tasks.In this paper, we introduce Sparse MetA-Tuning (SMAT), a method inspired bysparse mixture-of-experts approaches and trained to isolate subsets ofpre-trained parameters automatically for meta-tuning on each task. SMATsuccessfully overcomes OOD sensitivity and delivers on the promise of enhancingthe transfer abilities of vision foundation models beyond parameter-efficientfine-tuning. We establish new state-of-the-art results on a challengingcombination of Meta-Dataset augmented with additional OOD tasks in bothzero-shot and gradient-based adaptation settings. In addition, we provide athorough analysis of the superiority of learned over hand-designed sparsitypatterns for sparse expert methods and the pivotal importance of the sparsitylevel in balancing between in-distribution and out-of-distributiongeneralization. Our code is publicly available.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
few-shot-image-classification-on-meta-datasetSMAT (DINO-VIT-Base-16-224)
Accuracy: 85.27

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Unleashing the Power of Meta-tuning for Few-shot Generalization Through Sparse Interpolated Experts | Papers | HyperAI