HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation

Huiyu Wang Yukun Zhu Bradley Green Hartwig Adam Alan Yuille Liang-Chieh Chen

Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation

Abstract

Convolution exploits locality for efficiency at a cost of missing long range context. Self-attention has been adopted to augment CNNs with non-local interactions. Recent works prove it possible to stack self-attention layers to obtain a fully attentional network by restricting the attention to a local region. In this paper, we attempt to remove this constraint by factorizing 2D self-attention into two 1D self-attentions. This reduces computation complexity and allows performing attention within a larger or even global region. In companion, we also propose a position-sensitive self-attention design. Combining both yields our position-sensitive axial-attention layer, a novel building block that one could stack to form axial-attention models for image classification and dense prediction. We demonstrate the effectiveness of our model on four large-scale datasets. In particular, our model outperforms all existing stand-alone self-attention models on ImageNet. Our Axial-DeepLab improves 2.8% PQ over bottom-up state-of-the-art on COCO test-dev. This previous state-of-the-art is attained by our small variant that is 3.8x parameter-efficient and 27x computation-efficient. Axial-DeepLab also achieves state-of-the-art results on Mapillary Vistas and Cityscapes.

Code Repositories

MartinGer/Stand-Alone-Axial-Attention
pytorch
Mentioned in GitHub
csrhddlam/axial-deeplab
pytorch
Mentioned in GitHub
xiaofeng94/gmflownet
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
panoptic-segmentation-on-cityscapes-testAxial-DeepLab-XL (Mapillary Vistas, multi-scale)
PQ: 66.6
panoptic-segmentation-on-cityscapes-valAxial-DeepLab-XL (Mapillary Vistas, multi-scale)
AP: 44.2
PQ: 68.5
mIoU: 84.6
panoptic-segmentation-on-coco-minivalAxial-DeepLab-L (multi-scale)
PQ: 43.9
panoptic-segmentation-on-coco-minivalAxial-DeepLab-L(multi-scale)
PQst: 36.8
PQth: 48.6
panoptic-segmentation-on-coco-minivalAxial-DeepLab-L (single-scale)
PQ: 43.4
PQst: 35.6
PQth: 48.5
panoptic-segmentation-on-coco-test-devAxial-DeepLab-L (multi-scale)
PQ: 44.2
PQst: 36.8
PQth: 49.2
panoptic-segmentation-on-coco-test-devAxial-DeepLab-L
PQ: 43.6
PQst: 35.6
PQth: 48.9
panoptic-segmentation-on-mapillary-valAxial-DeepLab-L (multi-scale)
PQ: 41.1
PQst: 51.3
PQth: 33.4
mIoU: 58.4

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation | Papers | HyperAI