Command Palette
Search for a command to run...
Dissecting the impact of different loss functions with gradient surgery
Hong Xuan Robert Pless

Abstract
Pair-wise loss is an approach to metric learning that learns a semantic embedding by optimizing a loss function that encourages images from the same semantic class to be mapped closer than images from different classes. The literature reports a large and growing set of variations of the pair-wise loss strategies. Here we decompose the gradient of these loss functions into components that relate to how they push the relative feature positions of the anchor-positive and anchor-negative pairs. This decomposition allows the unification of a large collection of current pair-wise loss functions. Additionally, explicitly constructing pair-wise gradient updates to separate out these effects gives insights into which have the biggest impact, and leads to a simple algorithm that beats the state of the art for image retrieval on the CAR, CUB and Stanford Online products datasets.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| metric-learning-on-cars196 | Gradient Surgery | R@1: 86.5 |
| metric-learning-on-cub-200-2011 | Gradient Surgery | R@1: 63.8 |
| metric-learning-on-in-shop-1 | Gradient Surgery | R@1: 92.21 |
| metric-learning-on-stanford-online-products-1 | Gradient Surgery | R@1: 82.3 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.