HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

RedNet: Residual Encoder-Decoder Network for indoor RGB-D Semantic Segmentation

Jindong Jiang; Lunan Zheng; Fei Luo; Zhijun Zhang

RedNet: Residual Encoder-Decoder Network for indoor RGB-D Semantic Segmentation

Abstract

Indoor semantic segmentation has always been a difficult task in computer vision. In this paper, we propose an RGB-D residual encoder-decoder architecture, named RedNet, for indoor RGB-D semantic segmentation. In RedNet, the residual module is applied to both the encoder and decoder as the basic building block, and the skip-connection is used to bypass the spatial feature between the encoder and decoder. In order to incorporate the depth information of the scene, a fusion structure is constructed, which makes inference on RGB image and depth image separately, and fuses their features over several layers. In order to efficiently optimize the network's parameters, we propose a `pyramid supervision' training scheme, which applies supervised learning over different layers in the decoder, to cope with the problem of gradients vanishing. Experiment results show that the proposed RedNet(ResNet-50) achieves a state-of-the-art mIoU accuracy of 47.8% on the SUN RGB-D benchmark dataset.

Benchmarks

BenchmarkMethodologyMetrics
semantic-segmentation-on-nyu-depth-v2RedNet
Mean IoU: 47.2%
semantic-segmentation-on-sun-rgbdTokenFusion (Ti)
Mean IoU: 47.8%
semantic-segmentation-on-thud-robotic-datasetRedNet
mIoU: 76.92

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
RedNet: Residual Encoder-Decoder Network for indoor RGB-D Semantic Segmentation | Papers | HyperAI