HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

TransDSSL: Transformer based Depth Estimation via Self-Supervised Learning

{Yukyung Choi Soomnim Hwang Namil Kim Jeongmin Shin Daechan Han}

Abstract

Recently, transformers have been widely adopted for various computer vision tasks and show promising results due to their ability to encode long-range spatial dependencies in an image effectively. However, very few studies on adopting transformers in self-supervised depth estimation have been conducted. When replacing the CNN architecture with the transformer in self-supervised learning of depth, we encounter several problems such as problematic multi-scale photometric loss function when used with transformers and, insuffcient ability to capture local details. In this paper, we propose an attention-based decoder module, Pixel-Wise Skip Attention (PWSA), to enhance fine details in feature maps while keeping global context from transformers. In addition, we propose utilizing self-distillation loss with single-scale photometric loss to alleviate the instability of transformer training by using correct training signals. We demonstrate that the proposed model performs accurate predictions on large objects and thin structures that require global context and local details. Our model achieves state-ofthe-art performance among the self-supervised monocular depth estimation methods on KITTI and DDAD benchmarks

Benchmarks

BenchmarkMethodologyMetrics
monocular-depth-estimation-on-ddadTransDSSL
RMSE: 14.350
RMSE log: 0.172
Sq Rel: 3.591
absolute relative error: 0.151
monocular-depth-estimation-on-kitti-eigen-1TransDSSL
Delta u003c 1.25: 0.906
Delta u003c 1.25^2: 0.967
Delta u003c 1.25^3: 0.984
Mono: O
RMSE: 4.321
RMSE log: 0.172
Sq Rel: 0.711
absolute relative error: 0.095

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
TransDSSL: Transformer based Depth Estimation via Self-Supervised Learning | Papers | HyperAI