Command Palette
Search for a command to run...
Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels
Erik Englesson Hossein Azizpour

Abstract
Prior works have found it beneficial to combine provably noise-robust loss functions e.g., mean absolute error (MAE) with standard categorical loss function e.g. cross entropy (CE) to improve their learnability. Here, we propose to use Jensen-Shannon divergence as a noise-robust loss function and show that it interestingly interpolate between CE and MAE with a controllable mixing parameter. Furthermore, we make a crucial observation that CE exhibit lower consistency around noisy data points. Based on this observation, we adopt a generalized version of the Jensen-Shannon divergence for multiple distributions to encourage consistency around data points. Using this loss function, we show state-of-the-art results on both synthetic (CIFAR), and real-world (e.g., WebVision) noise with varying noise rates.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-classification-on-mini-webvision-1-0 | GJS (ResNet-50) | ImageNet Top-1 Accuracy: 75.50 ImageNet Top-5 Accuracy: 91.27 Top-1 Accuracy: 79.28 Top-5 Accuracy: 91.22 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.