HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Temporal Memory Attention for Video Semantic Segmentation

Hao Wang Weining Wang Jing Liu

Temporal Memory Attention for Video Semantic Segmentation

Abstract

Video semantic segmentation requires to utilize the complex temporal relations between frames of the video sequence. Previous works usually exploit accurate optical flow to leverage the temporal relations, which suffer much from heavy computational cost. In this paper, we propose a Temporal Memory Attention Network (TMANet) to adaptively integrate the long-range temporal relations over the video sequence based on the self-attention mechanism without exhaustive optical flow prediction. Specially, we construct a memory using several past frames to store the temporal information of the current frame. We then propose a temporal memory attention module to capture the relation between the current frame and the memory to enhance the representation of the current frame. Our method achieves new state-of-the-art performances on two challenging video semantic segmentation datasets, particularly 80.3% mIoU on Cityscapes and 76.5% mIoU on CamVid with ResNet-50.

Code Repositories

wanghao9610/TMANet
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
semantic-segmentation-on-urbanlfTMANet
mIoU (Real): 77.14
mIoU (Syn): 76.41
video-semantic-segmentation-on-camvidNetwarp
Mean IoU: 74.7
video-semantic-segmentation-on-camvidTMANet-50
Mean IoU: 76.5
video-semantic-segmentation-on-cityscapes-valTMANet-50
mIoU: 80.3

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Temporal Memory Attention for Video Semantic Segmentation | Papers | HyperAI