Command Palette
Search for a command to run...
Christopher Fifty Dennis Duan Ronald G. Junkins Ehsan Amid Jure Leskovec Christopher Re Sebastian Thrun

Abstract
Large Language Models like ChatGPT demonstrate a remarkable capacity to learn new concepts during inference without any fine-tuning. However, visual models trained to detect new objects during inference have been unable to replicate this ability, and instead either perform poorly or require meta-training and/or fine-tuning on similar objects. In this work, we propose a meta-learning algorithm that emulates Large Language Models by learning new visual concepts during inference without fine-tuning. Our approach leverages a frozen pre-trained feature extractor, and analogous to in-context learning, recasts visual meta-learning as sequence modeling over datapoints with known labels and a test datapoint with an unknown label. On 8 out of 11 meta-learning benchmarks, our approach -- without meta-training or fine-tuning -- exceeds or matches the state-of-the-art algorithm, P>M>F, which is meta-trained on these benchmarks. Our code is available at https://github.com/cfifty/CAML.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| few-shot-image-classification-on-cifar-fs-5 | CAML [Laion-2b] | Accuracy: 83.3 |
| few-shot-image-classification-on-cifar-fs-5-1 | CAML [Laion-2b] | Accuracy: 93.5 |
| few-shot-image-classification-on-cub-200-5 | CAML [Laion-2b] | Accuracy: 98.7 |
| few-shot-image-classification-on-cub-200-5-1 | CAML [Laion-2b] | Accuracy: 95.8 |
| few-shot-image-classification-on-mini-2 | CAML [Laion-2b] | Accuracy: 96.2 |
| few-shot-image-classification-on-mini-3 | CAML [Laion-2b] | Accuracy: 98.6 |
| few-shot-image-classification-on-tiered | CAML [Laion-2b] | Accuracy: 96.8 |
| few-shot-image-classification-on-tiered-1 | CAML [Laion-2b] | Accuracy: 98.8 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.