HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

DCPI-Depth: Explicitly Infusing Dense Correspondence Prior to Unsupervised Monocular Depth Estimation

Mengtan Zhang Yi Feng Qijun Chen Rui Fan

DCPI-Depth: Explicitly Infusing Dense Correspondence Prior to Unsupervised Monocular Depth Estimation

Abstract

There has been a recent surge of interest in learning to perceive depth from monocular videos in an unsupervised fashion. A key challenge in this field is achieving robust and accurate depth estimation in challenging scenarios, particularly in regions with weak textures or where dynamic objects are present. This study makes three major contributions by delving deeply into dense correspondence priors to provide existing frameworks with explicit geometric constraints. The first novelty is a contextual-geometric depth consistency loss, which employs depth maps triangulated from dense correspondences based on estimated ego-motion to guide the learning of depth perception from contextual information, since explicitly triangulated depth maps capture accurate relative distances among pixels. The second novelty arises from the observation that there exists an explicit, deducible relationship between optical flow divergence and depth gradient. A differential property correlation loss is, therefore, designed to refine depth estimation with a specific emphasis on local variations. The third novelty is a bidirectional stream co-adjustment strategy that enhances the interaction between rigid and optical flows, encouraging the former towards more accurate correspondence and making the latter more adaptable across various scenarios under the static scene hypotheses. DCPI-Depth, a framework that incorporates all these innovative components and couples two bidirectional and collaborative streams, achieves state-of-the-art performance and generalizability across multiple public datasets, outperforming all existing prior arts. Specifically, it demonstrates accurate depth estimation in texture-less and dynamic regions, and shows more reasonable smoothness. Our source code will be publicly available at mias.group/DCPI-Depth upon publication.

Benchmarks

BenchmarkMethodologyMetrics
monocular-depth-estimation-on-kitti-eigen-1DCPI-Depth (M+832x256+SC-V3)
RMSE: 4.496
Sq Rel: 0.679
absolute relative error: 0.109
monocular-depth-estimation-on-kitti-eigen-1DCPI-Depth (M+640x192)
Delta u003c 1.25: 0.902
Delta u003c 1.25^2: 0.967
Delta u003c 1.25^3: 0.985
RMSE: 4.274
RMSE log: 0.170
Sq Rel: 0.662
absolute relative error: 0.095
monocular-depth-estimation-on-kitti-eigen-1DCPI-Depth (M+1024x320)
Delta u003c 1.25: 0.914
Delta u003c 1.25^2: 0.969
Delta u003c 1.25^3: 0.985
Mono: O
RMSE: 4.113
RMSE log: 0.167
Sq Rel: 0.655
absolute relative error: 0.090

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
DCPI-Depth: Explicitly Infusing Dense Correspondence Prior to Unsupervised Monocular Depth Estimation | Papers | HyperAI