Command Palette
Search for a command to run...
{ Mehrtash Harandi Richard Nock Piotr Koniusz Christian Simon}

Abstract
Object recognition requires a generalization capability to avoid overfitting, especially when the samples are extremely few. Generalization from limited samples, usually studied under the umbrella of meta-learning, equips learning techniques with the ability to adapt quickly in dynamical environments and proves to be an essential aspect of life long learning. In this paper, we provide a framework for few-shot learning by introducing dynamic classifiers that are constructed from few samples. A subspace method is exploited as the central block of a dynamic classifier. We will empirically show that such modelling leads to robustness against perturbations (e.g., outliers) and yields competitive results on the task of supervised and semi-supervised few-shot classification. We also develop a discriminative form which can boost the accuracy even further. Our code is available at https://github.com/chrysts/dsn_fewshot
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| few-shot-image-classification-on-cifar-fs-5 | Adaptive Subspace Network | Accuracy: 78 |
| few-shot-image-classification-on-cifar-fs-5-1 | Adaptive Subspace Network | Accuracy: 87.3 |
| few-shot-image-classification-on-mini-2 | Adaptive Subspace Network | Accuracy: 67.09 |
| few-shot-image-classification-on-mini-3 | Adaptive Subspace Network | Accuracy: 81.65 |
| few-shot-image-classification-on-tiered | Adaptive Subspace Network | Accuracy: 68.44 |
| few-shot-image-classification-on-tiered-1 | Adaptive Subspace Network | Accuracy: 83.32 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.